Replies: 1 comment
-
If you want faster training speed, u can try to set the |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Trying to replicate SAM training from scratch.
I notice that during training the GPU usage goes up and down 0-100 even when not logging info. I am not computing embedding on the fly as I am loading the already precomputed.
I wonder if it's related to the grad accumulation step or something else?
I also see training is slower than sd2.1 in comparison even if I can use a bigger batch
Any insights?
Beta Was this translation helpful? Give feedback.
All reactions