Replies: 1 comment
-
@haimat we don't assist with code customizations but you might visit Multi-GPU Training tutorial for DDP examples commands. YOLOv5 Tutorials
Good luck 🍀 and let us know if you have any other questions! |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
We use the callback system in the YOLOv5 training function
train()
fromtrain.py
file in order to integrate it into our custom training frontend. This works fine, even when using Multi-GPU DataParallel Mode, running in one process. Now we would like to train in Multi-GPU DistributedDataParallel Mode, however, as far as I understand this would spawn multiple processes, right? So if this is the case, how can one train in DistributedDataParallel mode but still make use of the callbacks in the training script in order to get current status updates during training?Beta Was this translation helpful? Give feedback.
All reactions