Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

GPU Memory Requirement on Training MELO #2

Open
junsiknss opened this issue Jan 30, 2024 · 2 comments
Open

GPU Memory Requirement on Training MELO #2

junsiknss opened this issue Jan 30, 2024 · 2 comments

Comments

@junsiknss
Copy link

junsiknss commented Jan 30, 2024

Hello. I just read your paper.
In the paper, it is mentioned that the extra parameters are only needed ~0.2% (0.12M) of the original model (T5 small: 60M) when inferencing, but I didn't find anything about memory usage when training MELO.

Is it possible to get a rough idea of how much GPU Memory resources are required when training MELO?
Or if I'm misunderstanding the paper, please let me know.

Thanks.

@junsiknss junsiknss changed the title VRAM Requirement on Training MELO GPU Memory Requirement on Training MELO Jan 30, 2024
@BruthYU
Copy link
Member

BruthYU commented Mar 10, 2024

All experiments could be conducted on a single RTX 3090. I'll release some running logs recording the correct training and inference processes.

@BruthYU
Copy link
Member

BruthYU commented Mar 10, 2024

Logs could be downloaded using Google Drive. Thanks for your attention, please feel free to contact us whenever you have other questions.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants