Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

about training memory optimization #41

Closed
zhangvia opened this issue Dec 21, 2023 · 3 comments
Closed

about training memory optimization #41

zhangvia opened this issue Dec 21, 2023 · 3 comments

Comments

@zhangvia
Copy link

In the README, you mentioned that you would optimize the training code using DeepSpeed and Accelerate. However, as far as I know, the DeepSpeed functionality integrated into the Accelerate library does not support multi-model training. Do you have any suggestions?

@zhangvia
Copy link
Author

and From my test results, it appears that when resolution was set to 512, first-stage training cannot be conducted on a 40GB A100 even with a batch size of 1. Is this normal?

@guoqincode
Copy link
Owner

This is normal, 40G is too small for training.

@zhangvia
Copy link
Author

This is normal, 40G is too small for training.

what about the first question?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants