-
Notifications
You must be signed in to change notification settings - Fork 542
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
DistributedDataParallel #26
Comments
You can use |
I finally make it work with the help of diffdist, which provides a differentiable all_gather wrapper. |
Hi, can you share your code about how to implement this? I am not familiar with |
First, install diffdist.
|
Thank you for your quick reply. So, then I can simply compute the loss as usual and then backward the gradient? |
Yes, but I'm not sure this is bug-free.
…On Tue, Jul 21, 2020, 5:27 PM XU Shijian ***@***.***> wrote:
Thank you for your quick reply.
So, then I can simply compute the loss as usual and then backward the
gradient?
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
<#26 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AEFQBDVHNFCQAFCRFJGVORDR4VNO7ANCNFSM4O6IJ54A>
.
|
OK. Anyway, thanks a lot. |
Just for reference, this seems to be a reliable solution. |
Is there a way to make SupContrast work for DistributedDataParallel? By default each worker can only see its own sub-batch so the inter-sub-batch relationship of the samples will be utilized.
The text was updated successfully, but these errors were encountered: