File tree Expand file tree Collapse file tree 2 files changed +4
-8
lines changed
Expand file tree Collapse file tree 2 files changed +4
-8
lines changed Original file line number Diff line number Diff line change 3939
4040from distutils .util import strtobool
4141
42- from paddleformers .peft import LoRAModel , PrefixModelForCausalLM
42+ from paddleformers .peft import LoRAModel
4343from paddleformers .trainer import (
4444 speed_metrics ,
4545)
@@ -833,9 +833,7 @@ def fused_allreduce_gradients_no_sync(paramlist, hcg):
833833 logger .info (
834834 f"Loading best model from { self .state .best_model_checkpoint } (score: { self .state .best_metric } )."
835835 )
836- if isinstance (self .model , LoRAModel ) or isinstance (
837- self .model , PrefixModelForCausalLM
838- ):
836+ if isinstance (self .model , LoRAModel ):
839837 self ._load_best_model_from_peft_checkpoint ()
840838 else :
841839 weight_name = PADDLE_WEIGHTS_NAME
Original file line number Diff line number Diff line change 4141
4242from setuptools ._distutils .util import strtobool
4343
44- from paddleformers .peft import LoRAModel , PrefixModelForCausalLM
44+ from paddleformers .peft import LoRAModel
4545from paddleformers .trainer import (
4646 speed_metrics ,
4747)
@@ -829,9 +829,7 @@ def fused_allreduce_gradients_no_sync(paramlist, hcg):
829829 logger .info (
830830 f"Loading best model from { self .state .best_model_checkpoint } (score: { self .state .best_metric } )."
831831 )
832- if isinstance (self .model , LoRAModel ) or isinstance (
833- self .model , PrefixModelForCausalLM
834- ):
832+ if isinstance (self .model , LoRAModel ):
835833 self ._load_best_model_from_peft_checkpoint ()
836834 else :
837835 weight_name = PADDLE_WEIGHTS_NAME
You can’t perform that action at this time.
0 commit comments