You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: src/main.py
+3-6Lines changed: 3 additions & 6 deletions
Original file line number
Diff line number
Diff line change
@@ -117,10 +117,6 @@ def run(cfg):
117
117
# if a list, all see the exact splits of teams.
118
118
# if individual, they see different teams in splits. But as we show the average results, no big deal, esp., as we do n-fold
119
119
models= {}
120
-
# model names t* will follow the streaming scenario
121
-
# model names *_ts have timestamp (year) as a single added feature
122
-
# model names *_ts2v learn temporal skill vectors via d2v when each doc is a stream of (skills: year of the team)
123
-
# non-temporal (no streaming scenario, bag of teams)
124
120
assertlen(cfg.models.instances) >0, f'{opentf.textcolor["red"]}No model instance for training! Check ./src/__config__.yaml and models.instances ... {opentf.textcolor["reset"]}'
125
121
126
122
# Get command-line overrides for models. Kinda tricky as we dynamically override a subconfig.
assertskill_vecs.shape[0] ==teamsvecs['skill'].shape[0], f'{opentf.textcolor["red"]}Incorrect number of embeddings for teams subset of skills!{opentf.textcolor["reset"]}'
148
-
teamsvecs['original_skill'] =teamsvecs['skill'] #to accomodate skill_coverage metric and future use cases like in nmt
144
+
teamsvecs['original_skill'] =teamsvecs['skill'] #to accommodate skill_coverage metric and future use cases like in nmt
149
145
teamsvecs['skill'] =skill_vecs
150
146
151
147
formincfg.models.instances:
@@ -189,8 +185,9 @@ def run(cfg):
189
185
log.info(f'{opentf.textcolor["green"]}Aggregating the test results under {cfg.data.output} per splits from test.pred.eval.mean.csv files ... {opentf.textcolor["reset"]}')
190
186
aggregate(cfg.data.output)
191
187
192
-
# sample runs for different configs, including different prep, embeddings, model training, ..., are available as unit-test in
188
+
# sample runs for different configs, including different prep, embeddings, model training, ..., see unit-tests and scripts in
193
189
# ./github/workflows/*.yml
190
+
# ./ipynb/*.ipynb
194
191
195
192
# To run on compute canada servers you can use the following command: (time is in minutes)
0 commit comments