Skip to content

docs: clarify RL training GPU requirements by model config#764

Merged
felipemello1 merged 1 commit intometa-pytorch:mainfrom
kibitzing:docs/grpo-gpu-requirements
Feb 23, 2026
Merged

docs: clarify RL training GPU requirements by model config#764
felipemello1 merged 1 commit intometa-pytorch:mainfrom
kibitzing:docs/grpo-gpu-requirements

Conversation

@kibitzing
Copy link
Copy Markdown
Contributor

@kibitzing kibitzing commented Feb 21, 2026

  • I have personally reviewed this PR and description before asking others to do so. It meets the quality bar I expect from others. I understand that if this PR is perceived as unverified AI-generated code, it will be closed without further explanation.
  • I have run tests and confirmed that this code works

Description

The previous documentation stated a minimum requirement of 3+ GPUs for GRPO training. However, after #736, the reference model is no longer used by default.

As a result:

  • qwen3_1_7b now requires only 2 GPUs for RL training under the default configuration.
  • The minimum GPU requirement for RL training has been updated to 2 GPUs.

Additionally, the documentation listed 5 GPUs for Llama 3.1 8B. Based on the current YAML configurations, it requires 4 GPUs (generator + trainer), so the documentation has been corrected accordingly.

This PR updates the relevant documentation to reflect these changes.

Test plan

  • Manually verified GPU counts against YAML resource settings:
  • Manually reviewed updated markdown sections for consistency across pages.

@meta-cla meta-cla Bot added the CLA Signed This label is managed by the Meta Open Source bot. label Feb 21, 2026
@felipemello1 felipemello1 merged commit d3eb3bf into meta-pytorch:main Feb 23, 2026
6 checks passed
@felipemello1
Copy link
Copy Markdown
Contributor

ty!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

CLA Signed This label is managed by the Meta Open Source bot.

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants