docs: add LoRA fine-tuning tutorial#3601
docs: add LoRA fine-tuning tutorial#3601chiajunglien wants to merge 2 commits intoAI-Hypercomputer:jackyf/feat/lora-nnxfrom
Conversation
docs/tutorials/posttraining/lora.md
Outdated
| scan_layers=True | ||
| ``` | ||
|
|
||
| Your fine-tuned model checkpoints will be saved here: `$BASE_OUTPUT_DIRECTORY/$RUN_NAME/checkpoints`. No newline at end of file |
There was a problem hiding this comment.
Also instruct the usage of maxtext_lora_to_hf script
| learning_rate="${LEARNING_RATE?}" \ | ||
| weight_dtype="${WEIGHT_DTYPE?}" \ | ||
| dtype="${DTYPE?}" \ | ||
| profiler=xplane \ |
There was a problem hiding this comment.
I think we don't need to use profiler in tutorial
|
|
||
| Your fine-tuned model checkpoints will be saved here: `$BASE_OUTPUT_DIRECTORY/$RUN_NAME/checkpoints`. | ||
|
|
||
| ## (Optional) Export Fine-tuned LoRA to Hugging Face Format |
There was a problem hiding this comment.
Convert should be more appropriate
|
|
||
| ```sh | ||
| python3 maxtext/checkpoint_conversion/maxtext_to_hf_lora.py \ | ||
| maxtext/configs/post_train/sft.yml \ |
There was a problem hiding this comment.
Could we remove maxtext/configs/post_train/sft.yml?
|
|
||
| ```sh | ||
| python3 maxtext/checkpoint_conversion/hf_lora_to_maxtext.py \ | ||
| maxtext/configs/post_train/sft.yml \ |
There was a problem hiding this comment.
Could we remove maxtext/configs/post_train/sft.yml?
| maxtext/configs/post_train/sft.yml \ | ||
| model_name="${PRE_TRAINED_MODEL?}" \ | ||
| load_parameters_path="${BASE_OUTPUT_DIRECTORY?}/${RUN_NAME?}/checkpoints/<step_number>/items" \ | ||
| base_output_directory="${BASE_OUTPUT_DIRECTORY?}/hf_lora_adaptor" \ |
There was a problem hiding this comment.
Rest of the file use adapter and we should align
| If your LoRA adapter is currently in Hugging Face format, you must convert it to MaxText format before it can be loaded. Use the provided conversion script: | ||
|
|
||
| ```sh | ||
| python3 maxtext/checkpoint_conversion/hf_lora_to_maxtext.py \ |
There was a problem hiding this comment.
we should use python3 -m maxtext.checkpoint_conversion.hf_lora_to_maxtext to align the training command
| After completing the fine-tuning process, your LoRA weights are stored in MaxText/Orbax format. To use these weights with the Hugging Face ecosystem (e.g., for inference or sharing), convert them back using the `maxtext_lora_to_hf.py` script. | ||
|
|
||
| ```sh | ||
| python3 maxtext/checkpoint_conversion/maxtext_to_hf_lora.py \ |
There was a problem hiding this comment.
we should use python3 -m maxtext.checkpoint_conversion.maxtext_to_hf_lora to align the training command
Description
Start with a short description of what the PR does and how this is a change from
the past.
The rest of the description includes relevant details and context, examples:
If the change fixes a bug or a Github issue, please include a link, e.g.,:
FIXES: b/123456
FIXES: #123456
Notice 1: Once all tests pass, the "pull ready" label will automatically be assigned.
This label is used for administrative purposes. Please do not add it manually.
Notice 2: For external contributions, our settings currently require an approval from a MaxText maintainer to trigger CI tests.
Tests
Please describe how you tested this change, and include any instructions and/or
commands to reproduce.
Checklist
Before submitting this PR, please make sure (put X in square brackets):
gemini-reviewlabel.