-
Notifications
You must be signed in to change notification settings - Fork 476
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Yi-34B 需要的资源是多少? #55
Comments
@xgysigned 4090/3090 的显存应该在 24 GB,34B 参数以 float16/bfloat16 加载需要 34 GB*2=68 GB 左右显存,需要上多卡 多卡可以考虑使用仓库中的 TP Demo:https://github.com/01-ai/Yi/blob/main/demo/text_generation_tp.py 也可以考虑使用 vllm/llamacpp 等社区开源的推理框架的一些特性来进一步降低显存的需求以及提升推理性能: |
可以考虑支持 |
希望量化后的版本支持24G显卡。 |
希望量化后的版本支持24G显卡。 |
卡一個量化後版本 |
4块2080ti魔改22g的显卡(22g*4=88g)可以跑吗?目前我有两块 |
4bit量化版本应该正好支持24G显存的显卡 |
q4_k_s is suitable for deploying on Tesla P40 (24G VRAM). |
yi-34b需要的资源是多少?单卡3090或4090能跑吗?还是需要多卡?
The text was updated successfully, but these errors were encountered: