-
Notifications
You must be signed in to change notification settings - Fork 7.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Feature] Add llama 3 model #2239
Comments
for me it works, but there is an issue: after the first answer the end of the answer does not seem to be detected, CPU stays at 100%... |
Yes very much hoping for Llama3 in GPT4all! |
This fixes the issue for me with GGUFs: Problem: Llama-3 uses 2 different stop tokens, but llama.cpp only has support for one. The instruct models seem to always generate a <|eot_id|> but the GGUF uses <|end_of_text|>. Solution: Edit the GGUF file so it uses the correct stop token.
|
Thank you, very helpful! I'm on an M1 Macbook Air with 16GB Ram. I downloaded this model |
Maybe I spoke too soon, in another test it kept talking and didn't stop until I told it to. So maybe I need to mess with the correct stop token too. |
No, I think then you need to change the max number of tokens you want, or manually press stop |
fixed instruct model link : https://huggingface.co/bartowski/Meta-Llama-3-8B-Instruct-GGUF |
Feature Request
Are we going to support llama 3 model: https://github.com/meta-llama/llama3
The text was updated successfully, but these errors were encountered: