Skip to content

Conversation

@HollowMan6
Copy link
Collaborator

@HollowMan6 HollowMan6 commented Dec 22, 2025

What does this PR do?

Waiting for:

lora-performance lora-critic-val-score lora-actor-plus-rollout-mismatch

Checklist Before Starting

  • Search for similar PRs. Paste at least one query link here: ...
  • Format the PR title as [{modules}] {type}: {description} (This will be checked by the CI)
    • {modules} include fsdp, megatron, sglang, vllm, rollout, trainer, ci, training_utils, recipe, hardware, deployment, ray, worker, single_controller, misc, perf, model, algo, env, tool, ckpt, doc, data, cfg, reward
    • If this PR involves multiple modules, separate them with , like [megatron, fsdp, doc]
    • {type} is in feat, fix, refactor, chore, test
    • If this PR breaks any API (CLI arguments, config, function signature, etc.), add [BREAKING] to the beginning of the title.
    • Example: [BREAKING][fsdp, megatron] feat: dynamic batching

Test

For changes that can not be tested by CI (e.g., algorithm implementation, new model support), validate by experiment(s) and show results like training curve plots, evaluation results, etc.

API and Usage Example

Demonstrate how the API changes if any, and provide usage example(s) if possible.

# Add code snippet or script demonstrating how to use this

Design & Code Changes

Demonstrate the high-level design if this PR is complex, and list the specific changes.

Checklist Before Submitting

Important

Please check all the following items before requesting a review, otherwise the reviewer might deprioritize this PR for review.

@HollowMan6 HollowMan6 changed the title [megatron] feat: LoRA adapter only weight update [megatron] feat: LoRA adapter only weight update (TensorLoRARequest) Dec 22, 2025
@HollowMan6 HollowMan6 changed the title [megatron] feat: LoRA adapter only weight update (TensorLoRARequest) [megatron] feat: LoRA adapter only refit (TensorLoRARequest) Dec 22, 2025
Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request introduces a significant feature for LoRA adapter-only weight updates in the Megatron backend, which avoids the overhead of merging adapters into the base model for each weight synchronization. The changes are well-structured and include updates to documentation, configuration handling for LoRA, new PEFT utility functions for vLLM compatibility, and refactored weight export logic. The implementation appears solid and aligns with the stated goals. I have a couple of suggestions to enhance code maintainability and clarity regarding duplicated logic and a confusing condition.

@HollowMan6 HollowMan6 force-pushed the lora_adapters_update branch 4 times, most recently from f6706bc to 81d261e Compare December 22, 2025 01:33
@HollowMan6 HollowMan6 force-pushed the lora_adapters_update branch 5 times, most recently from 23b8716 to 797dc7f Compare December 28, 2025 14:01
Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request introduces a valuable feature for LoRA adapter-only refitting in the Megatron backend, which should provide significant performance benefits. The changes across documentation, configuration, and worker implementations appear to correctly support both merging LoRA adapters and loading them separately. My primary concern, detailed in a specific comment, is the repeated implementation of LoRA configuration logic across several files. Addressing this will improve the long-term maintainability of the codebase.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant