|
1 |
| -# Config for multi-device LoRA finetuning in lora_finetune_distributed_td.py |
| 1 | +# Config for multi-device LoRA finetuning in lora_finetune_distributed_multi_dataset.py |
2 | 2 | # using a Llama3.2 11B Vision Instruct model
|
3 | 3 | #
|
4 | 4 | # This config assumes that you've run the following command before launching:
|
5 | 5 | # tune download meta-llama/Llama-3.2-11B-Vision-Instruct --output-dir /tmp/Llama-3.2-11B-Vision-Instruct --ignore-patterns "original/consolidated*"
|
6 | 6 | #
|
7 | 7 | # To launch on 2 devices, run the following command from root:
|
8 |
| -# tune run --nproc_per_node 2 lora_finetune_distributed_td --config llama3_2_vision/11B_lora_td |
| 8 | +# tune run --nproc_per_node 2 lora_finetune_distributed_multi_dataset --config llama3_2_vision/11B_lora_multi_dataset |
9 | 9 | #
|
10 | 10 | # You can add specific overrides through the command line. For example
|
11 | 11 | # to override the checkpointer directory while launching training:
|
12 |
| -# tune run --nproc_per_node 2 lora_finetune_distributed_td --config llama3_2_vision/11B_lora_td checkpointer.checkpoint_dir=<YOUR_CHECKPOINT_DIR> |
| 12 | +# tune run --nproc_per_node 2 lora_finetune_distributed_multi_dataset --config llama3_2_vision/11B_lora_multi_dataset checkpointer.checkpoint_dir=<YOUR_CHECKPOINT_DIR> |
13 | 13 | #
|
14 | 14 | # This config works best when the model is being fine-tuned on 2+ GPUs.
|
15 | 15 | # For single device LoRA finetuning please use 11B_lora_single_device.yaml
|
16 | 16 | # or 11B_qlora_single_device.yaml
|
17 | 17 |
|
| 18 | +output_dir: /tmp/torchtune/llama3_2_vision_11B/lora_multi_dataset # /tmp may be deleted by your system. Change it to your preference. |
| 19 | + |
18 | 20 | # Model arguments
|
19 | 21 | model:
|
20 | 22 | _component_: torchtune.models.llama3_2_vision.lora_llama3_2_vision_11b
|
@@ -44,7 +46,7 @@ checkpointer:
|
44 | 46 | filename_format: model-{}-of-{}.safetensors
|
45 | 47 | max_filename: "00005"
|
46 | 48 | recipe_checkpoint: null
|
47 |
| - output_dir: /tmp/Llama-3.2-11B-Vision-Instruct/ |
| 49 | + output_dir: ${output_dir} |
48 | 50 | model_type: LLAMA3_VISION
|
49 | 51 | resume_from_checkpoint: False
|
50 | 52 | save_adapter_weights_only: False # PeFT formatting not available yet. This will save it in torchtune format only.
|
@@ -117,6 +119,6 @@ dtype: bf16
|
117 | 119 | output_dir: /tmp/lora-llama3.2-vision-finetune
|
118 | 120 | metric_logger:
|
119 | 121 | _component_: torchtune.training.metric_logging.DiskLogger
|
120 |
| - log_dir: /tmp/Llama-3.2-11B-Vision-Instruct/logs |
| 122 | + log_dir: ${output_dir}/logs |
121 | 123 | log_every_n_steps: 1
|
122 | 124 | log_peak_memory_stats: True
|
0 commit comments