File tree 1 file changed +4
-8
lines changed
1 file changed +4
-8
lines changed Original file line number Diff line number Diff line change 14
14
import torch .distributed as dist
15
15
from torch import nn
16
16
from torch ._C ._distributed_c10d import (
17
+ _resolve_process_group ,
17
18
AllgatherOptions ,
18
19
AllreduceOptions ,
19
20
BroadcastOptions ,
20
21
ReduceOp ,
21
- _resolve_process_group ,
22
22
)
23
23
from torch .distributed import (
24
+ _functional_collectives ,
25
+ get_world_size ,
24
26
ReduceOp ,
25
27
TCPStore ,
26
28
Work ,
27
- _functional_collectives ,
28
- get_world_size ,
29
29
)
30
30
from torch .distributed .device_mesh import init_device_mesh
31
31
from torch .testing ._internal .common_distributed import MultiProcessTestCase
34
34
from torchft .process_group import (
35
35
_DummyWork ,
36
36
_ErrorSwallowingWork ,
37
+ _ManagedWork ,
37
38
ErrorSwallowingProcessGroupWrapper ,
38
39
extend_device_mesh ,
39
40
ft_init_device_mesh ,
46
47
ProcessGroupGloo ,
47
48
ProcessGroupNCCL ,
48
49
ProcessGroupWrapper ,
49
- _DummyWork ,
50
- _ErrorSwallowingWork ,
51
- _ManagedWork ,
52
- extend_device_mesh ,
53
50
)
54
51
55
52
@@ -319,7 +316,6 @@ def setUp(self):
319
316
self ._spawn_processes ()
320
317
321
318
def test_init_device_mesh (self ) -> None :
322
- os .environ ["MASTER_ADDR" ] = "localhost"
323
319
os .environ ["MASTER_PORT" ] = str (12346 )
324
320
os .environ ["RANK" ] = str (self .rank )
325
321
os .environ ["WORLD_SIZE" ] = str (4 )
You can’t perform that action at this time.
0 commit comments