You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi Tengfei Wang, such a amazing reasearch and many thanks for sharing the code. Very intersting results...
I was able to reproduce some results and really liked the work flow you created of CNN and not Oflow, seams it handles perspective shifts and background better (still playing with it).
The dilate mask makes totally sense...
My question is about multi-GPU to speed up training....im doing these below:
on train.py i removed the # on mirrored_strategy = tf.distribute.MirroredStrategy() line
and
added # on os.environ["CUDA_VISIBLE_DEVICES"] = FLAGS.GPU_ID.
With that seams that is Training is using both GPUs, but also shows that the GPU_0 is using CUDA and processing but GPU_1 only using vram, does not seams to be using CUDA and process, only VRAM.
Is that correct?
Also saw @tf.function down below, but not sure if i should remove # on those lines. Also found #dist_full_ds = mirrored_strategy, tried but seams to do the same thing on second gpu, only using vram, not processing
Is that correct behavior?
Thank you Tengfei Wang and once again, amazing research.
The text was updated successfully, but these errors were encountered:
Hi, thanks for your interest in our work. We just add the distributed training code 'train_dist.py' in this repo. It may take a few minutes to initialize the distributed training when you run train_dist.py.
Currently, the distributed training code only works on tf 2.0, due to the API changes of TensorFlow.
Hi Tengfei Wang, such a amazing reasearch and many thanks for sharing the code. Very intersting results...
I was able to reproduce some results and really liked the work flow you created of CNN and not Oflow, seams it handles perspective shifts and background better (still playing with it).
The dilate mask makes totally sense...
My question is about multi-GPU to speed up training....im doing these below:
on train.py i removed the # on mirrored_strategy = tf.distribute.MirroredStrategy() line
and
added # on os.environ["CUDA_VISIBLE_DEVICES"] = FLAGS.GPU_ID.
With that seams that is Training is using both GPUs, but also shows that the GPU_0 is using CUDA and processing but GPU_1 only using vram, does not seams to be using CUDA and process, only VRAM.
Is that correct?
Also saw @tf.function down below, but not sure if i should remove # on those lines. Also found #dist_full_ds = mirrored_strategy, tried but seams to do the same thing on second gpu, only using vram, not processing
Is that correct behavior?
Thank you Tengfei Wang and once again, amazing research.
The text was updated successfully, but these errors were encountered: