You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+1-1
Original file line number
Diff line number
Diff line change
@@ -12,6 +12,6 @@ In their investigation the authors figure out:
12
12
* Different snapshots (i.e. model from epoch 1, model from epoch 2, and so on) of a same model exhibit functional similarity. Hence, their ensemble is **less likely** to explore the different modes of local minimas in the optimization space.
13
13
* Different solutions of a same model (i.e. trained with different random initializations each time) exhibit functional dissimilarity. Hence, their ensemble is **more likely** to explore the different modes of local minimas in the optimization space.
14
14
15
-
Along with this fascinating finding, they present a number of different things that are useful to understand the dynamics of deep neural networks in general. To know more about them check out [our report](https://app.wandb.ai/authors/loss-landscape/reports/Understanding-the-effectivity-of-ensembles-in-deep-learning-(tentative)--VmlldzoxODAxNjA).
15
+
Along with these fascinating findings, they present a number of different things that are useful to understand the dynamics of deep neural networks in general. To know more about them check out [our report](https://app.wandb.ai/authors/loss-landscape/reports/Understanding-the-effectivity-of-ensembles-in-deep-learning-(tentative)--VmlldzoxODAxNjA).
16
16
17
17
Huge to shoutout to **Yannic Kilcher** for making a stellar [explanation video of the paper](https://www.youtube.com/watch?v=5IRlUVrEVL8). It helped us accelerating our experimentation process.
0 commit comments