-
Notifications
You must be signed in to change notification settings - Fork 28
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
abnormal training time when run multi NeuRad jobs #32
Comments
Have you checked that the jobs do not use the same resources (GPU, CPU)? |
I see. We often train multiple jobs in parallel on our cluster as well and have never had any issues where they affect each other. I know that the multiprocess data loading has been given some people issues, not sure if that is the case here as well? |
Hi, the issue was solved by setting |
The training time becomes longer when I run the second job in a multi-GPU cluster.

And then, the second job's training time is also slower as below.

Could you give me some suggestions?
Thank you in advance.
The text was updated successfully, but these errors were encountered: