File tree 3 files changed +6
-5
lines changed
3 files changed +6
-5
lines changed Original file line number Diff line number Diff line change @@ -202,7 +202,7 @@ def test_fx_script(self) -> None:
202
202
sparse_features = sparse_features ,
203
203
)
204
204
205
- gm = symbolic_trace (deepfm_nn )
205
+ gm = torch . fx . GraphModule (deepfm_nn , Tracer (). trace ( deepfm_nn ) )
206
206
207
207
scripted_gm = torch .jit .script (gm )
208
208
Original file line number Diff line number Diff line change 11
11
from typing import Dict
12
12
13
13
import torch
14
+ from torchrec .fx import Tracer
14
15
from torchrec .modules .mc_modules import (
15
16
average_threshold_filter ,
16
17
DistanceLFU_EvictionPolicy ,
@@ -357,5 +358,5 @@ def test_fx_jit_script_not_training(self) -> None:
357
358
)
358
359
359
360
model .train (False )
360
- gm = torch .fx .symbolic_trace (model )
361
+ gm = torch .fx .GraphModule (model , Tracer (). trace ( model ) )
361
362
torch .jit .script (gm )
Original file line number Diff line number Diff line change 14
14
import torch
15
15
from hypothesis import given , settings
16
16
from torch import nn
17
- from torchrec .fx import symbolic_trace
17
+ from torchrec .fx import symbolic_trace , Tracer
18
18
from torchrec .modules .mlp import MLP , Perceptron
19
19
20
20
@@ -99,13 +99,13 @@ def test_fx_script_Perceptron(self) -> None:
99
99
# Dry-run to initialize lazy module.
100
100
m (torch .randn (batch_size , in_features ))
101
101
102
- gm = symbolic_trace ( m )
102
+ gm = torch . fx . GraphModule ( m , Tracer (). trace ( m ) )
103
103
torch .jit .script (gm )
104
104
105
105
def test_fx_script_MLP (self ) -> None :
106
106
in_features = 3
107
107
layer_sizes = [16 , 8 , 4 ]
108
108
m = MLP (in_features , layer_sizes )
109
109
110
- gm = symbolic_trace ( m )
110
+ gm = torch . fx . GraphModule ( m , Tracer (). trace ( m ) )
111
111
torch .jit .script (gm )
You can’t perform that action at this time.
0 commit comments