-
Notifications
You must be signed in to change notification settings - Fork 24
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
multi-training #11
Comments
Hi, i'm not sure what your question is. Isn't it natural that multi-GPU DDP training costs around the same time per iteration as single GPU training? |
Yes,The eta of multi-GPU is the same as single gpu for 17 days |
In the config we set the total number of training iterations, so changing the number of GPUs will only affect the total number of epochs but not the training time. And btw, the initial eta is an unreliable overestimate. |
I see.that means if i change the total_iters to be half in multi-GPU training,we can get the same result as single-GPU training?I remember only 6 days for multi-GPU training in your paper |
There's no need to change the schedule, the 17 days eta is simply wrong. |
okay,I see. |
If i use 4-gpus,can i change the total_iters to be half(500k) to get the simlar result as 1000k with 2-gpus? |
Why using your code multi graphics card training is as fast as single-card training (2x3090)?
The text was updated successfully, but these errors were encountered: