We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent 21b3922 commit 4739ce1Copy full SHA for 4739ce1
src/transformers/trainer.py
@@ -793,7 +793,7 @@ def train(
793
model = ShardedDDP(model, self.optimizer)
794
elif is_sagemaker_distributed_available():
795
model = DDP(model, device_ids=[dist.get_local_rank()], broadcast_buffers=False)
796
- if self.deepspeed:
+ elif self.deepspeed:
797
pass # already initialized its own DDP earlier
798
elif self.args.local_rank != -1:
799
if self.args.ddp_find_unused_parameters is not None:
0 commit comments