-
Notifications
You must be signed in to change notification settings - Fork 88
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
a multi-gpu bug #136
Comments
Could you provide a snippet of code that reproduces the problem? If possible, make one that doesn't need a multi-gpu setup to reproduce it, as I don't have such a setup available at the moment |
Ok, but without a code example and a multi-gpu setup I won't be able to reproduce the bug at the moment. Does this bug apply to all transforms, or just ApplyImpulseResponse and/or AddBackgroundNoise? Is there a way you can work around it? Like always give it a batch size that is divisible by your number of GPUs? A fixed batch size should do the trick. Or would you like to make a PR that fixes the bug? Or maybe I should slap a known limitation on readme that says multi-GPU with "uneven" batch sizes isn't officially supported? |
When I Use batch_sizes which can be divible by the number of GPUs, it can work normally. It seems like occur when there are uneven batch sizes in different GPUs. If you can't reproduce the bug and fix that, I think I can have a try to fix this bug by myself at first. |
If I use the batchbins in espnet, It will trigger a multi-GPU bug.
for example, if I use two GPUs, and the final batch_size is 61, and I use data parallel, it will divide into 30, 31,
when I thy to use torch-audiomentations, it will trigger a bug as follow.
whether The batch size of each card must be the same or there can be other solutions to avoid this bug.
looking forward to a reply
The text was updated successfully, but these errors were encountered: