Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Reduce training memory requirement #39

Open
LifeIsStrange opened this issue Dec 1, 2022 · 2 comments
Open

Reduce training memory requirement #39

LifeIsStrange opened this issue Dec 1, 2022 · 2 comments

Comments

@LifeIsStrange
Copy link

LifeIsStrange commented Dec 1, 2022

CUDA-enabled machine (48 GB to train, 4 GB to evaluate)

@vdobrovolskii friendly ping
Are 48GB really needed to train?
Can't we train longer (how long) with less ?
couldn't your project leverage FP16, FP8 and other optimizations ? You can get them out of the box if you use roberta from the Transformers library https://github.com/huggingface/transformers
Also there is accelerate https://huggingface.co/docs/accelerate/index

I have a 3070 with 8GB of GDDR6 :/

@vdobrovolskii
Copy link
Owner

It should be totally possible to reduce the training requirements, and I've been thinking a long time about rewriting the project (because it grew out of another project and there's a lot of legacy in it) using Pytorch Lightning to allow for easy access to the optimizations, multi-gpu training, etc.

I just haven't had time to do that yet :(

@LifeIsStrange
Copy link
Author

Happy to hear that :)
No worries you don't owe us anything but that would be great if you find the time/energy/will, plz ping me if that happen someday !

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants