Replies: 1 comment
-
|
If you want faster training speed, u can try to set the |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Trying to replicate SAM training from scratch.
I notice that during training the GPU usage goes up and down 0-100 even when not logging info. I am not computing embedding on the fly as I am loading the already precomputed.
I wonder if it's related to the grad accumulation step or something else?
I also see training is slower than sd2.1 in comparison even if I can use a bigger batch
Any insights?
Beta Was this translation helpful? Give feedback.
All reactions