New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
run infer failed #40
Comments
Could you please provide the error log? I can run the huggingface code on a single A100 GPU. |
first is OOM ,and I try a shorter prompt(问题:冬天中国...) ,it worked,but the result is not right, so I think maybe it's "half", so I delete "half", OOM again. |
Because GLM-10B is the English model. We will upload GLM-10B-Chinese in the future. |
We uploaded GLM-10B-Chinese and GLM-Large-Chinese in the huggingface hub. |
I use A100 40G * 8 to run the huggingface hub code and failed.
I try to add
device_map='auto'
at AutoModelForSeq2SeqLM.from_pretrained,but not support.how to run this code?
The text was updated successfully, but these errors were encountered: