Skip to content

Add Nemotron 3 Nano 30B multi-node training tutorial#699

Merged
bxyu-nvidia merged 14 commits intomainfrom
add-nemotron-3-nano-30b-tutorial
Feb 19, 2026
Merged

Add Nemotron 3 Nano 30B multi-node training tutorial#699
bxyu-nvidia merged 14 commits intomainfrom
add-nemotron-3-nano-30b-tutorial

Conversation

@srogawski-nvidia
Copy link
Copy Markdown
Contributor

@srogawski-nvidia srogawski-nvidia commented Feb 15, 2026

Summary

Adds a new tutorial for training Nemotron 3 Nano 30B on 32 nodes with GRPO, building on the existing multi-node training guide.

Changes

New tutorial: nemotron-3-nano-30b-multi-node.md for 32-node (256 GPU) training
Updated: multi-node-training.md to link to the new advanced tutorial

Testing

Tested with:

  • 2-node jobs (16 GPUs)
  • 32-node jobs (256 GPUs)
  • Both completed successfully with proper Ray cluster formation

Documentation Flow

Single Node Training -> Multi-Node Training -> Nemotron 3 Nano 30B (new) -> Custom Environment

Closes #389

Signed-off-by: Sebastian Rogawski <srogawski@nvidia.com>
@copy-pr-bot
Copy link
Copy Markdown

copy-pr-bot bot commented Feb 15, 2026

This pull request requires additional validation before any workflows can run on NVIDIA's runners.

Pull request vetters can view their responsibilities here.

Contributors can view more details about this message here.

Copy link
Copy Markdown
Contributor

@bxyu-nvidia bxyu-nvidia left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

this looks amazing, thank you so much! one minor request

```bash
# Set workspace directory (adjust to your cluster's large storage)
# Examples: /scratch/$USER, /work/$USER, /data/$USER, /lustre/.../users/$USER
export WORKSPACE=/path/to/large/storage/$USER
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

can we avoid exporting this and instead leave it as a local bash variable?

Copy link
Copy Markdown
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

removed export

Signed-off-by: Sebastian Rogawski <srogawski@nvidia.com>
Signed-off-by: Sebastian Rogawski <srogawski@nvidia.com>
@srogawski-nvidia srogawski-nvidia force-pushed the add-nemotron-3-nano-30b-tutorial branch from b82cd51 to 600316a Compare February 16, 2026 00:38
@cwing-nvidia cwing-nvidia self-requested a review February 16, 2026 02:56
@cwing-nvidia
Copy link
Copy Markdown
Contributor

Could we move the tutorial to a new toctree section Model Recipes? I think that will improve discoverability

image

@lbliii can you help offer guidance on tutorial header formatting (goals, duration, pre-reqs etc)

Comment thread docs/training-tutorials/nemo-rl-grpo/nemotron-3-nano-30b-multi-node.md Outdated
Comment thread docs/training-tutorials/nemo-rl-grpo/nemotron-3-nano-30b-multi-node.md Outdated
Comment thread docs/training-tutorials/nemo-rl-grpo/nemotron-3-nano-30b-multi-node.md Outdated
Comment thread docs/model-recipes/nemotron-3-nano.md
Comment thread docs/model-recipes/nemotron-3-nano.md
Comment thread docs/training-tutorials/nemo-rl-grpo/nemotron-3-nano-30b-multi-node.md Outdated
Comment thread docs/training-tutorials/nemo-rl-grpo/nemotron-3-nano-30b-multi-node.md Outdated
Signed-off-by: Sebastian Rogawski <srogawski@nvidia.com>
…rocessing, safetensors verification

Signed-off-by: Sebastian Rogawski <srogawski@nvidia.com>
@srogawski-nvidia srogawski-nvidia force-pushed the add-nemotron-3-nano-30b-tutorial branch from 95904d1 to 51f7037 Compare February 17, 2026 05:25
@srogawski-nvidia srogawski-nvidia force-pushed the add-nemotron-3-nano-30b-tutorial branch from f6db5a3 to 15e7293 Compare February 17, 2026 16:47
…0b-tutorial

Signed-off-by: Sebastian Rogawski <srogawski@nvidia.com>
Comment thread docs/training-tutorials/nemo-rl-grpo/nemotron-3-nano-30b-multi-node.md Outdated
@bxyu-nvidia
Copy link
Copy Markdown
Contributor

re: Move just the new tutorial doc location from docs/training-tutorials/nemo-rl-grpo/nemotron-3-nano-30b-multi-node.md -> docs/model-recipes/nemotron-3-nano.md

For links, we can keep the Workplace Assistant doc holistically as a prerequisite, just frame it as the entire tutorial

Comment thread docs/training-tutorials/nemo-rl-grpo/nemotron-3-nano-30b-multi-node.md Outdated
…d GitHub links

Signed-off-by: Sebastian Rogawski <srogawski@nvidia.com>
…0b-tutorial

Signed-off-by: Sebastian Rogawski <srogawski@nvidia.com>
@srogawski-nvidia srogawski-nvidia force-pushed the add-nemotron-3-nano-30b-tutorial branch from feb7332 to 8e4b717 Compare February 18, 2026 00:00
…tainer.sh) to avoid the subshell export issue

Signed-off-by: Sebastian Rogawski <srogawski@nvidia.com>
Signed-off-by: Sebastian Rogawski <srogawski@nvidia.com>
@bxyu-nvidia bxyu-nvidia merged commit 5a5b6e1 into main Feb 19, 2026
5 checks passed
@bxyu-nvidia bxyu-nvidia deleted the add-nemotron-3-nano-30b-tutorial branch February 19, 2026 04:23
fsiino-nvidia pushed a commit that referenced this pull request Feb 21, 2026
## Summary
Adds a new tutorial for training Nemotron 3 Nano 30B on 32 nodes with
GRPO, building on the existing multi-node training guide.

## Changes
**New tutorial**: `nemotron-3-nano-30b-multi-node.md` for 32-node (256
GPU) training
**Updated**: `multi-node-training.md` to link to the new advanced
tutorial

## Testing
Tested with:
- 2-node jobs (16 GPUs)
- 32-node jobs (256 GPUs)
- Both completed successfully with proper Ray cluster formation

## Documentation Flow
Single Node Training -> Multi-Node Training -> **Nemotron 3 Nano 30B**
(new) -> Custom Environment

 Closes #389

---------

Signed-off-by: Sebastian Rogawski <srogawski@nvidia.com>
Signed-off-by: Brian Yu <bxyu@nvidia.com>
Co-authored-by: bxyu-nvidia <bxyu@nvidia.com>
fsiino-nvidia pushed a commit that referenced this pull request Feb 21, 2026
## Summary
Adds a new tutorial for training Nemotron 3 Nano 30B on 32 nodes with
GRPO, building on the existing multi-node training guide.

## Changes
**New tutorial**: `nemotron-3-nano-30b-multi-node.md` for 32-node (256
GPU) training
**Updated**: `multi-node-training.md` to link to the new advanced
tutorial

## Testing
Tested with:
- 2-node jobs (16 GPUs)
- 32-node jobs (256 GPUs)
- Both completed successfully with proper Ray cluster formation

## Documentation Flow
Single Node Training -> Multi-Node Training -> **Nemotron 3 Nano 30B**
(new) -> Custom Environment

 Closes #389

---------

Signed-off-by: Sebastian Rogawski <srogawski@nvidia.com>
Signed-off-by: Brian Yu <bxyu@nvidia.com>
Co-authored-by: bxyu-nvidia <bxyu@nvidia.com>
fsiino-nvidia pushed a commit that referenced this pull request Feb 21, 2026
## Summary
Adds a new tutorial for training Nemotron 3 Nano 30B on 32 nodes with
GRPO, building on the existing multi-node training guide.

## Changes
**New tutorial**: `nemotron-3-nano-30b-multi-node.md` for 32-node (256
GPU) training
**Updated**: `multi-node-training.md` to link to the new advanced
tutorial

## Testing
Tested with:
- 2-node jobs (16 GPUs)
- 32-node jobs (256 GPUs)
- Both completed successfully with proper Ray cluster formation

## Documentation Flow
Single Node Training -> Multi-Node Training -> **Nemotron 3 Nano 30B**
(new) -> Custom Environment

 Closes #389

---------

Signed-off-by: Sebastian Rogawski <srogawski@nvidia.com>
Signed-off-by: Brian Yu <bxyu@nvidia.com>
Co-authored-by: bxyu-nvidia <bxyu@nvidia.com>
Signed-off-by: Frankie Siino <fsiino@nvidia.com>
abubakaria56 pushed a commit to abubakaria56/Gym that referenced this pull request Mar 2, 2026
## Summary
Adds a new tutorial for training Nemotron 3 Nano 30B on 32 nodes with
GRPO, building on the existing multi-node training guide.

## Changes
**New tutorial**: `nemotron-3-nano-30b-multi-node.md` for 32-node (256
GPU) training
**Updated**: `multi-node-training.md` to link to the new advanced
tutorial

## Testing
Tested with:
- 2-node jobs (16 GPUs)
- 32-node jobs (256 GPUs)
- Both completed successfully with proper Ray cluster formation

## Documentation Flow
Single Node Training -> Multi-Node Training -> **Nemotron 3 Nano 30B**
(new) -> Custom Environment

 Closes NVIDIA-NeMo#389

---------

Signed-off-by: Sebastian Rogawski <srogawski@nvidia.com>
Signed-off-by: Brian Yu <bxyu@nvidia.com>
Co-authored-by: bxyu-nvidia <bxyu@nvidia.com>
abubakaria56 pushed a commit to abubakaria56/Gym that referenced this pull request Mar 2, 2026
## Summary
Adds a new tutorial for training Nemotron 3 Nano 30B on 32 nodes with
GRPO, building on the existing multi-node training guide.

## Changes
**New tutorial**: `nemotron-3-nano-30b-multi-node.md` for 32-node (256
GPU) training
**Updated**: `multi-node-training.md` to link to the new advanced
tutorial

## Testing
Tested with:
- 2-node jobs (16 GPUs)
- 32-node jobs (256 GPUs)
- Both completed successfully with proper Ray cluster formation

## Documentation Flow
Single Node Training -> Multi-Node Training -> **Nemotron 3 Nano 30B**
(new) -> Custom Environment

 Closes NVIDIA-NeMo#389

---------

Signed-off-by: Sebastian Rogawski <srogawski@nvidia.com>
Signed-off-by: Brian Yu <bxyu@nvidia.com>
Co-authored-by: bxyu-nvidia <bxyu@nvidia.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

Nemotron 3 Nano recipe release pointer

4 participants