You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi , I dont use much pytorch-lightning or performed distributed training (noob here),
but found that the training never starts on a single gpu single node configuration
The solution which I found out for the same was to modify the num_nodes parameter in your train configuration to 1
If number is greater that 1 pytorch lightning waits for the other nodes I presume
It took a lot of time for me to get it right , putting it out for fellow noobs : )
Thanks for sharing such an incredible work to the community !!!
The text was updated successfully, but these errors were encountered:
Hi , I dont use much pytorch-lightning or performed distributed training (noob here),
but found that the training never starts on a single gpu single node configuration
The solution which I found out for the same was to modify the
num_nodes
parameter in your train configuration to 1If number is greater that 1 pytorch lightning waits for the other nodes I presume
It took a lot of time for me to get it right , putting it out for fellow noobs : )
Thanks for sharing such an incredible work to the community !!!
The text was updated successfully, but these errors were encountered: