-
Notifications
You must be signed in to change notification settings - Fork 2
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
model: Mistral Nemo #19
Comments
@nguyenhoangthuan99 Can you look into this:
|
|
@offgridtech I am transferring this issue to cortex.cpp repo. We should be working on it, ETA 2 weeks |
I think Mistral Nemo is the first model for us to do this pipeline automatically. To add new model support from hugging face
|
Mistral-nemo is supported now at cortexso. 10 quantization levels are available now. All models are created and uploaded automatically through CI. |
closing as done and QA'd |
Current behavior
I see a bunch of stuff on HuggingFace and llama.cpp Git about pre-tokenizers causing issues upon initial release of the quantizied Mistal Nemo model, but it seemed everything was cleared up over the last few days due to a llama.cpp update. What worked for other people didn't work for Jan. I've tried several quant versions, and it fails to start. Saw KoboldCPP and LMStudio say they made some updates, and it's fixed now. I'm guessing you all need to do the same. Thanks
More information here:
ggerganov/llama.cpp#8579
ggerganov/llama.cpp#8604
Minimum reproduction step
It doesn't start. Other models like llama 3.1 start fine.
Expected behavior
The model starts
Screenshots / Logs
This log looks like it is the pre-tokenizer issue they were talking about.
Jan version
v0.5.2
In which operating systems have you tested?
Environment details
AppImage on Linux
The text was updated successfully, but these errors were encountered: