-
Notifications
You must be signed in to change notification settings - Fork 14
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
About multi-gpu training #2
Comments
Hi, thanks for having a look at the code. I did not test dual-gpu training, and RN101 indeed takes quite some time on single GPU (~2 weeks). I did not do the effort of implementing multi-gpu support, since I had to use the other available GPUs in our lab for other runs/experiments. I also plan to release a trained mobilenetv2 with the optimized CUDA code integrated. |
Hi, @thomasverelst Furthermore, I have made attempts towards multi-gpu training by simply wrapping the model with
Looking forward to your good news! Also congratulations on the upcoming MobileNetV2 CUDA code! |
I've pushed a new branch |
Yeah, it seems to work now. I have successfully run this branch with ResNet-32 on CIFAR for fast prototyping (with matched accuracy and reduced FLOPs). As an additional note, the "FLOPs counting to zero" problem can be solved by modifying the following line |
Thanks a lot, that fixed it. |
Thanks for your awesome work! Is there any idea how multi-gpu training is supported? Because you know training ResNet-101 on ImageNet with a single GPU is unacceptably slow.
The text was updated successfully, but these errors were encountered: