Replies: 2 comments
-
|
>>> erogol |
Beta Was this translation helpful? Give feedback.
0 replies
-
|
>>> alchemi5t |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
>>> alchemi5t
[September 5, 2019, 7:16am]
I've been trying to use multiple gpus for training but the training
hangs(Responsive, but indefinitely stays here without aby output or
error) at initializing process group. Any insights on how to fix this?
Tried pytorch 0.4.1 and 1+.
'distributed':{
'backend': 'nccl',
'url': 'tcp: slash / slash /localhost:23456'
},
[This is an archived TTS discussion thread from discourse.mozilla.org/t/hangs-on-dist-init-process-group-in-distribute-py]
Beta Was this translation helpful? Give feedback.
All reactions