Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Same amount of VRAM is taken as in AdamW #22

Open
VCasecnikovs opened this issue Apr 3, 2023 · 6 comments
Open

Same amount of VRAM is taken as in AdamW #22

VCasecnikovs opened this issue Apr 3, 2023 · 6 comments

Comments

@VCasecnikovs
Copy link

One of the main benefits of LION, is it needs to save less data for each param.
Adam needs to save Momentum and RMSProp ema's, while in LION we need to save only momentum ema.
When I try to use LION, it takes exactly the same amount of memory as AdamW

@xiangning-chen
Copy link
Contributor

Hi, what is the model size in your setting?
When the model is small, I think the main memory overhead comes from the activation, so the saved second moment may not be significant.

@VCasecnikovs
Copy link
Author

@xiangning-chen
178m parameters, convolutional.

@feffy380
Copy link

feffy380 commented Apr 7, 2023

Are you comparing this to AdamW8bit by chance?

@VCasecnikovs
Copy link
Author

No, to AdamW

@konev-artem
Copy link

In my setting, Lion takes less memory than AdamW (9.9 Gb vs 10.1Gb) but Lion is slower in terms of steps/sec. Has anyone noticed the same? I compare Lion with triton vs fused AdamW.

@nicosouth
Copy link

do you solve the problem? i have the same problem.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants