You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am retraining AA-RMVSNet in DTU dataset with default settings on one V100 32GB GPU.
But it cost about 17GB for batchsize=1, and batchsize=2 will cause OOM problem.
It is really strange because in the paper, batchsize=4 costs only 20.16GB. Besides, the depth_num is set as 192 in the paper, while it is just 150 in the default setting.
Another question is that the training is very slow. It cost about 4.6s for one step of batch=1.
Can you provide any advice on it?
The text was updated successfully, but these errors were encountered:
Hello. As illustrated in the paper, we use 4 GPUs for training, and the resulting total batchsize is 4. Since you are attempting to put a batch of 4 on only one GPU, it is natural to get OOM. You can adjust the depth division according to the hint in train_dtu.sh by modifying both d and interval_scale.
And in terms of training speed, since RNN has to process one slice by another, your training speed seems right to me.
Thanks for your great works!
I am retraining AA-RMVSNet in DTU dataset with default settings on one V100 32GB GPU.
But it cost about 17GB for batchsize=1, and batchsize=2 will cause OOM problem.
It is really strange because in the paper, batchsize=4 costs only 20.16GB. Besides, the depth_num is set as 192 in the paper, while it is just 150 in the default setting.
Another question is that the training is very slow. It cost about 4.6s for one step of batch=1.
Can you provide any advice on it?
The text was updated successfully, but these errors were encountered: