Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Can you support DeepSeek's inference acceleration? Thank you very much. #351

Open
joyhhheee opened this issue Dec 7, 2023 · 4 comments
Open

Comments

@joyhhheee
Copy link

No description provided.

@mrwyattii
Copy link
Contributor

Hi @joyhhheee could you be a little more specific? Are you referring to supporting the models that DeepSeek has shared on HuggingFace?

@lynquantumman
Copy link

Run deepseek model on mii will generate repetitive results, don't if something got wrong

@brianyu-nexusflowai
Copy link

@ys2899
Copy link

ys2899 commented Apr 5, 2024

The same problem here, I zoomed in and look at the codes why it generates repetitive results. I locate the problems over
orig_module.forward = inference_container.module.forward
in hybrid engine.
Anybody has insights about this?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants