-
Notifications
You must be signed in to change notification settings - Fork 42
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Can we just use the sloth gradient checkpointing by uncommenting this line? #30
Comments
Yes you can. It will produce the same loss. But it does not enable greater batch size in my experiments. |
I am getting this error though when I do this. Any idea why ? |
I need it reduce memory footprint not batch size |
A question assert gradient_checkpointing_kwargs == None is there which throws an error. Do I need to set gradient_checkpointing_kwargs to something or I need to comment this line? |
I was not clear about how to use the code ?
https://github.com/jzhang38/EasyContext/blob/main/train.py#L28 By uncommenting this line we can enable sloth code?
The text was updated successfully, but these errors were encountered: