Skip to content

Commit 8421d6b

Browse files
authored
fix: Update Nemotron Nano v3 Configs (#1122)
Signed-off-by: Robert Clark <roclark@nvidia.com>
1 parent 5809056 commit 8421d6b

File tree

2 files changed

+10
-4
lines changed

2 files changed

+10
-4
lines changed

examples/llm_finetune/nemotron/nemotron_nano_v3_squad.yaml

Lines changed: 5 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -13,8 +13,10 @@
1313
# limitations under the License.
1414

1515

16-
# To run this recipe, please use the following command:
17-
# torchrun --nproc-per-node=8 recipes/llm_finetune/finetune.py --config recipes/llm_finetune/nemotron/llama3_3_nemotron_super_49B_squad_peft.yaml
16+
# To run this recipe, please use the following command. Replace --nproc-per-node=8 with the number of GPUs on your node:
17+
# torchrun --nproc-per-node=8 examples/llm_finetune/finetune.py --config examples/llm_finetune/nemotron/nemotron_nano_v3_squad.yaml
18+
# For multi-node, use the following:
19+
# torchrun --nproc-per-node=8 --nnodes=2 examples/llm_finetune/finetune.py --config examples/llm_finetune/nemotron/nemotron_nano_v3_squad.yaml
1820

1921

2022
step_scheduler:
@@ -36,6 +38,7 @@ rng:
3638
model:
3739
_target_: nemo_automodel.NeMoAutoModelForCausalLM.from_pretrained
3840
pretrained_model_name_or_path: nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B-BF16
41+
trust_remote_code: true
3942

4043
# torch.compile configuration
4144
compile:

examples/llm_finetune/nemotron/nemotron_nano_v3_squad_peft.yaml

Lines changed: 5 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -13,8 +13,10 @@
1313
# limitations under the License.
1414

1515

16-
# To run this recipe, please use the following command:
17-
# torchrun --nproc-per-node=8 recipes/llm_finetune/finetune.py --config recipes/llm_finetune/nemotron/llama3_3_nemotron_super_49B_squad_peft.yaml
16+
# To run this recipe, please use the following command. Replace --nproc-per-node=8 with the number of GPUs on your node:
17+
# torchrun --nproc-per-node=8 examples/llm_finetune/finetune.py --config examples/llm_finetune/nemotron/nemotron_nano_v3_squad_peft.yaml
18+
# For multi-node, use the following:
19+
# torchrun --nproc-per-node=8 --nnodes=2 examples/llm_finetune/finetune.py --config examples/llm_finetune/nemotron/nemotron_nano_v3_squad_peft.yaml
1820

1921

2022
step_scheduler:
@@ -36,6 +38,7 @@ rng:
3638
model:
3739
_target_: nemo_automodel.NeMoAutoModelForCausalLM.from_pretrained
3840
pretrained_model_name_or_path: nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B-BF16
41+
trust_remote_code: true
3942

4043
# torch.compile configuration
4144
compile:

0 commit comments

Comments
 (0)