Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

how much gpus did you use? #48

Closed
Tord-Zhang opened this issue Sep 15, 2020 · 4 comments
Closed

how much gpus did you use? #48

Tord-Zhang opened this issue Sep 15, 2020 · 4 comments

Comments

@Tord-Zhang
Copy link

Hi, since the MidasNet is a very large model, how much gpus did you use and how long did it take to train the model? Since the batchsize is not larget ( 8 for each dataset), would multi-gpu training hurt the performance? Since there are a lot of batch normalization layers in the encoder.

Thanks.

@ranftlr
Copy link
Collaborator

ranftlr commented Sep 15, 2020

The model was trained on 2 Quadro RTX 6000s for about 6 days: 1 day for pre-training on RedWeb + about 5 days for training on MIX5. Multi-GPU with low batch size was not an issue, because we froze all batch norm layers.

@Tord-Zhang
Copy link
Author

@ranftlr thanks for your quick responce. I am training MidasNet on 2 tesla v100 gpus, but it's much slower. So the mean and val of bn layers is not updated during the training process?

@ranftlr
Copy link
Collaborator

ranftlr commented Sep 16, 2020

Correct, mean and variance are fixed and are not updated. I can't comment on the speed on a v100 as I don't have any available.

@ranftlr ranftlr closed this as completed Oct 20, 2020
@mathmax12
Copy link

@Tord-Zhang
I am reproducing the training processing on one V100. I am pretty new to this. I wonder what's the amount of data and how long the training takes?
Thanks

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants