Skip to content

Commit 045b398

Browse files
committed
Test
1 parent 2db49d3 commit 045b398

File tree

1 file changed

+6
-0
lines changed

1 file changed

+6
-0
lines changed

tests/test_trainer_distributed.py

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -118,16 +118,22 @@ def test_gaudi_trainer_distributed_hpu_graphs(self):
118118
#
119119
# PYTHONPATH="src" python optimum-habana/examples/gaudi_spawn.py --world_size 8 --use_mpi --output_dir output_dir ./tests/test_trainer_distributed.py
120120

121+
print("000")
121122
parser = HfArgumentParser((GaudiTrainingArguments,))
123+
print("111")
122124
training_args = parser.parse_args_into_dataclasses()[0]
125+
print("222")
123126

124127
gaudi_config_file = Path(__file__).parent.resolve() / Path("configs/gaudi_config_trainer_test.json")
128+
print("333")
125129
gaudi_config = GaudiConfig.from_pretrained(gaudi_config_file)
130+
print("444")
126131

127132
logger.warning(
128133
f"Process rank: {training_args.local_rank}, device: {training_args.device}, n_hpu: {training_args.world_size},"
129134
f" distributed training: {training_args.local_rank != -1}"
130135
)
136+
print("555")
131137

132138
# Essentially, what we want to verify in the distributed case is that we get all samples back,
133139
# in the right order. (this is crucial for prediction for instance)

0 commit comments

Comments
 (0)