Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

run infer failed #40

Closed
xv44586 opened this issue Nov 2, 2022 · 4 comments
Closed

run infer failed #40

xv44586 opened this issue Nov 2, 2022 · 4 comments

Comments

@xv44586
Copy link
Contributor

xv44586 commented Nov 2, 2022

I use A100 40G * 8 to run the huggingface hub code and failed.
I try to add device_map='auto' at AutoModelForSeq2SeqLM.from_pretrained,but not support.
how to run this code?

@duzx16
Copy link
Member

duzx16 commented Nov 2, 2022

Could you please provide the error log? I can run the huggingface code on a single A100 GPU.

@xv44586
Copy link
Contributor Author

xv44586 commented Nov 4, 2022

first is OOM ,and I try a shorter prompt(问题:冬天中国...) ,it worked,but the result is not right, so I think maybe it's "half", so I delete "half", OOM again.
the result is not right because is the "half" or I need use the Chinese version?

@duzx16
Copy link
Member

duzx16 commented Nov 4, 2022

first is OOM ,and I try a shorter prompt(问题:冬天中国...) ,it worked,but the result is not right, so I think maybe it's "half", so I delete "half", OOM again. the result is not right because is the "half" or I need use the Chinese version?

Because GLM-10B is the English model. We will upload GLM-10B-Chinese in the future.

@duzx16
Copy link
Member

duzx16 commented Nov 19, 2022

We uploaded GLM-10B-Chinese and GLM-Large-Chinese in the huggingface hub.

@duzx16 duzx16 closed this as completed Nov 19, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants