Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to load model with load_8_bit #24

Closed
stevezhang88 opened this issue Apr 18, 2023 · 3 comments
Closed

How to load model with load_8_bit #24

stevezhang88 opened this issue Apr 18, 2023 · 3 comments

Comments

@stevezhang88
Copy link

My GPU is 3090ti, 24G. I have to use load_8_bit to load vicuna 13b. Could you can tell me how to do on MiniGPT-4?

@TsuTikgiau
Copy link
Collaborator

Thank you for your interest! Currently, you can set low_resource to True in the evaluation config file eval_configs/minigpt4_eval.yaml. Besides, you may need to set the number of beams to 1 to save GPU memory in demo.py.
As discussed in [this issue](Another user in this issue, this reduces the memory usage to about 24G, although it may still OOM when the output is long. We are working to come up with a solution to make it work inside 24G memory. Will come back to you once we finish

@stevezhang88
Copy link
Author

thanks. you're so nice. waiting for your solution. I suppose many person like me have only one-GPU computer.

@TsuTikgiau
Copy link
Collaborator

We update the default setting of the demo and now it should load vicuna in 8 bit by default. The demo now should be able to run on single 3090 if you set the beam search width to 1(which is also the default value now)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants