@@ -90,7 +90,7 @@ You do not need to enter the Docker container. Just set the config and run.
9090EXP=examples/megatron/configs/MI300X/llama3.1_8B-pretrain.yaml bash ./examples/run_local_pretrain.sh
9191
9292# examples for torchtitan llama3.1_8B
93- EXP=examples/torchtitan/mi300 /llama3.1_8B-pretrain.yaml bash ./examples/run_local_pretrain.sh
93+ EXP=examples/torchtitan/configs/MI300X /llama3.1_8B-pretrain.yaml bash ./examples/run_local_pretrain.sh
9494```
9595
9696---
@@ -114,7 +114,7 @@ cd Primus && pip install -r requirements.txt
114114EXP=examples/megatron/configs/MI300X/llama3.1_8B-pretrain.yaml bash ./examples/run_pretrain.sh
115115
116116# examples for torchtitan llama3.1_8B
117- EXP=examples/torchtitan/mi300 /llama3.1_8B-pretrain.yaml bash ./examples/run_pretrain.sh
117+ EXP=examples/torchtitan/configs/MI300X /llama3.1_8B-pretrain.yaml bash ./examples/run_pretrain.sh
118118
119119```
120120
@@ -133,7 +133,7 @@ export NNODES=8
133133EXP=examples/megatron/configs/MI300X/llama3.1_8B-pretrain.yaml bash ./examples/run_slurm_pretrain.sh
134134
135135# examples for torchtitan llama3.1_8b
136- EXP=examples/torchtitan/mi300 /llama3.1_8B-pretrain.yaml bash ./examples/run_slurm_pretrain.sh
136+ EXP=examples/torchtitan/configs/MI300X /llama3.1_8B-pretrain.yaml bash ./examples/run_slurm_pretrain.sh
137137```
138138
139139## 🔧 HipblasLT Auto Tuning
@@ -199,9 +199,9 @@ The following models are supported out of the box via provided configuration fil
199199| llama2_70B | [ meta-llama/Llama-2-70b-hf] ( https://huggingface.co/meta-llama/Llama-2-70b-hf ) | [ llama2_70B-pretrain.yaml] ( https://github.com/AMD-AIG-AIMA/Primus/blob/main/examples/megatron/configs/MI300X/llama2_70B-pretrain.yaml ) | |
200200| llama3_8B | [ meta-llama/Meta-Llama-3-8B] ( https://huggingface.co/meta-llama/Meta-Llama-3-8B ) | [ llama3_8B-pretrain.yaml] ( https://github.com/AMD-AIG-AIMA/Primus/blob/main/examples/megatron/configs/MI300X/llama3_8B-pretrain.yaml ) | |
201201| llama3_70B | [ meta-llama/Meta-Llama-3-70B] ( https://huggingface.co/meta-llama/Meta-Llama-3-70B ) | [ llama3_70B-pretrain.yaml] ( https://github.com/AMD-AIG-AIMA/Primus/blob/main/examples/megatron/configs/MI300X/llama3_70B-pretrain.yaml ) | |
202- | llama3.1_8B | [ meta-llama/Llama-3.1-8B] ( https://huggingface.co/meta-llama/Llama-3.1-8B ) | [ llama3.1_8B-pretrain.yaml] ( https://github.com/AMD-AIG-AIMA/Primus/blob/main/examples/megatron/configs/MI300X/llama3.1_8B-pretrain.yaml ) | [ llama3.1_8B-pretrain.yaml] ( https://github.com/AMD-AIG-AIMA/Primus/blob/main/examples/torchtitan/mi300 /llama3.1_8B-pretrain.yaml ) |
203- | llama3.1_70B | [ meta-llama/Llama-3.1-70B] ( https://huggingface.co/meta-llama/Llama-3.1-70B ) | [ llama3.1_70B-pretrain.yaml] ( https://github.com/AMD-AIG-AIMA/Primus/blob/main/examples/megatron/configs/MI300X/llama3.1_70B-pretrain.yaml ) | [ llama3.1_70B-pretrain.yaml] ( https://github.com/AMD-AIG-AIMA/Primus/blob/main/examples/torchtitan/mi300 /llama3.1_70B-pretrain.yaml ) |
204- | llama3.1_405B | [ meta-llama/Llama-3.1-405B] ( https://huggingface.co/meta-llama/Llama-3.1-405B ) | [ llama3.1_405B-pretrain.yaml] ( https://github.com/AMD-AIG-AIMA/Primus/blob/main/examples/megatron/configs/MI300X/llama3.1_405B-pretrain.yaml ) | [ llama3.1_405B-pretrain.yaml] ( https://github.com/AMD-AIG-AIMA/Primus/blob/main/examples/torchtitan/mi300 /llama3.1_405B-pretrain.yaml ) |
202+ | llama3.1_8B | [ meta-llama/Llama-3.1-8B] ( https://huggingface.co/meta-llama/Llama-3.1-8B ) | [ llama3.1_8B-pretrain.yaml] ( https://github.com/AMD-AIG-AIMA/Primus/blob/main/examples/megatron/configs/MI300X/llama3.1_8B-pretrain.yaml ) | [ llama3.1_8B-pretrain.yaml] ( https://github.com/AMD-AIG-AIMA/Primus/blob/main/examples/torchtitan/configs/MI300X /llama3.1_8B-pretrain.yaml ) |
203+ | llama3.1_70B | [ meta-llama/Llama-3.1-70B] ( https://huggingface.co/meta-llama/Llama-3.1-70B ) | [ llama3.1_70B-pretrain.yaml] ( https://github.com/AMD-AIG-AIMA/Primus/blob/main/examples/megatron/configs/MI300X/llama3.1_70B-pretrain.yaml ) | [ llama3.1_70B-pretrain.yaml] ( https://github.com/AMD-AIG-AIMA/Primus/blob/main/examples/torchtitan/configs/MI300X /llama3.1_70B-pretrain.yaml ) |
204+ | llama3.1_405B | [ meta-llama/Llama-3.1-405B] ( https://huggingface.co/meta-llama/Llama-3.1-405B ) | [ llama3.1_405B-pretrain.yaml] ( https://github.com/AMD-AIG-AIMA/Primus/blob/main/examples/megatron/configs/MI300X/llama3.1_405B-pretrain.yaml ) | [ llama3.1_405B-pretrain.yaml] ( https://github.com/AMD-AIG-AIMA/Primus/blob/main/examples/torchtitan/configs/MI300X /llama3.1_405B-pretrain.yaml ) |
205205| deepseek_v2_lite | [ deepseek-ai/DeepSeek-V2-Lite] ( https://huggingface.co/deepseek-ai/DeepSeek-V2-Lite ) | [ deepseek_v2_lite-pretrain.yaml] ( https://github.com/AMD-AIG-AIMA/Primus/blob/main/examples/megatron/configs/MI300X/deepseek_v2_lite-pretrain.yaml ) | |
206206| deepseek_v2 | [ deepseek-ai/DeepSeek-V2] ( https://huggingface.co/deepseek-ai/DeepSeek-V2 ) | [ deepseek_v2-pretrain.yaml] ( https://github.com/AMD-AIG-AIMA/Primus/blob/main/examples/megatron/configs/MI300X/deepseek_v2-pretrain.yaml ) | |
207207| deepseek_v3 | [ deepseek-ai/DeepSeek-V3] ( https://huggingface.co/deepseek-ai/DeepSeek-V3 ) | [ deepseek_v3-pretrain.yaml] ( https://github.com/AMD-AIG-AIMA/Primus/blob/main/examples/megatron/configs/MI300X/deepseek_v3-pretrain.yaml ) | |
@@ -223,7 +223,7 @@ For example, to run the llama3.1_8B model quickly:
223223``` bash
224224EXP=examples/megatron/configs/MI300X/llama3.1_8B-pretrain.yaml bash ./examples/run_local_pretrain.sh
225225
226- EXP=examples/torchtitan/mi300 /llama3.1_8B-pretrain.yaml bash ./examples/run_local_pretrain.sh
226+ EXP=examples/torchtitan/configs/MI300X /llama3.1_8B-pretrain.yaml bash ./examples/run_local_pretrain.sh
227227```
228228
229229
@@ -236,7 +236,7 @@ export NNODES=8
236236EXP=examples/megatron/configs/MI300X/llama3.1_8B-pretrain.yaml bash ./examples/run_slurm_pretrain.sh
237237
238238# run torchtitan
239- EXP=examples/torchtitan/mi300 /llama3.1_8B-pretrain.yaml bash ./examples/run_slurm_pretrain.sh
239+ EXP=examples/torchtitan/configs/MI300X /llama3.1_8B-pretrain.yaml bash ./examples/run_slurm_pretrain.sh
240240```
241241
242242## ☸️ Kubernetes Training Management (` run_k8s_pretrain.sh ` )
0 commit comments