-
Notifications
You must be signed in to change notification settings - Fork 382
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
GPU memory and speed #4
Comments
With the default setup, it takes ~24GB GPU memory and ~16 hours to train 500k iterations on A100 GPUs. There are many hyperparameters that can affect performance (both GPU memory and time). The most significant one is probably the hash table size (
These would likely come with the cost of slightly lower quality in the results though. |
Hi @mowangmodi Thank you for your interest in the project. As @chenhsuanlin mentioned, both dictionary size and batch size will affect the training requirements. Another pointer to reduce memory usage is the |
Closing due to inactivity, please feel free to reopen if there are further issues! |
Please see the FAQ section on how to adjust the hyperparameters |
|
@chenhsuanlin Thanks for your effort on this great project first. I have some questions.
#57 (comment) |
How much GPU memory is required for training and how much time is spent
The text was updated successfully, but these errors were encountered: