Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Gpu memory exceeded when training a a model #25

Closed
pyz-creeper opened this issue Mar 6, 2023 · 2 comments
Closed

Gpu memory exceeded when training a a model #25

pyz-creeper opened this issue Mar 6, 2023 · 2 comments

Comments

@pyz-creeper
Copy link

Hi, thanks for the great work. I was experimenting on an adversarial training method using your multi-modality model, but the gpu memory ran out very quickly. Can you share roughly how much gpu memory is required to train the model assuming the batch size is one?

@yanwei-li
Copy link
Member

yanwei-li commented Mar 9, 2023

Hi, thanks for your interest. For the largest model LiDAR-V0075-R101, I used the A100, it may cost about 64G with Batch size 2. That means about 32G if the batch size is set to 1.
If you do not need to train it from scratch, it's better to fix the pre-trained model, which reduces the cost a lot.

@pyz-creeper
Copy link
Author

Hi, thanks for your interest. For the largest model LiDAR-V0075-R101, I used the A100, it may cost about 64G with Batch size 2. That means about 32G if the batch size is set to 1. If you do not need to train it from scratch, it's better to fix the pre-trained model, which reduces the cost a lot.

Thanks for your reply, I am going to try it with a bigger gpu.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants