Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Quick question about training time and compute #11

Closed
greeneggsandyaml opened this issue Mar 10, 2023 · 2 comments
Closed

Quick question about training time and compute #11

greeneggsandyaml opened this issue Mar 10, 2023 · 2 comments

Comments

@greeneggsandyaml
Copy link

Hello, thanks for your great work. I have a very question about training.

I'm trying to run training and getting an OutOfMemoryError using a (single) 32 GB GPU (V100). What do you use for training? Also, with your compute setup, approximately how long does training take?

Thanks so much!

@shariqfarooq123
Copy link
Collaborator

shariqfarooq123 commented Mar 10, 2023

Hi, thanks for appreciating our work!

For metric fine-tuning, We use 4 NVIDIA A100 GPUs for training our largest model (BEiT-L). Training time on NYU (~25k samples, 5 epochs) on 4 A100s (40GB) is less than 2 hours.

Relative pre-training on 12 datasets (M12 from the paper) takes around 3-5 days on 8 RTX A6000-like GPUs. This gives us the MiDaS v3.1 models. Please refer to midas repo for more details.

@greeneggsandyaml
Copy link
Author

Brilliant, thanks for the quick response. It's great to hear that the training time is quick.

I'll create another issue if I have any more questions, but this is resolved, so I'm closing the issue.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants