New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Is wizard 2 working with llamacpp? #6691
Comments
it should work. They're a family of finetune based on mixtral and mistral7b |
Yes, it works. I think this is how to run it in
|
This only generates "##########" for me! (a bunch of pound signs) |
Your prompt ends with an end-of-sequence token and provides no new sequence for the model to begin with. Your in-prefix and in-suffix also don't have the newline characters required by the Vicuna prompt format. I ran a couple queries using Mikupad and the llama.cpp server and the imatrix quant I pulled from here worked just fine with proper Vicuna formatting. |
@4onen Worked for me. Anyway, there's no newlines between the end of input and beggining of Assistant in your link. It's consistent with vicuna prompt template, showing seperation by a space then a stop token.
@21stcaveman I'm not sure what caused pound symbols. |
My mistake on the prompt formatting. Visually it looks like interactive mode is adding newlines there, but I don't have much experience with local interactive mode because I want to be able to go back and edit both my inputs and the AI output, which last time I checked interactive mode didn't support. |
Can confirm the original example by @Jeximo works. Problem seems to have been the GGUF copy I had. Download from the link provided by @4onen fixed the issue. |
Is wizard 2 working with llamacp?
On the paper looks insane... better than gpt-4 , mistal large , claudie 3 sonet, etc
The text was updated successfully, but these errors were encountered: