-
Notifications
You must be signed in to change notification settings - Fork 8.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Llama3-8b & Perplexity.exe Issue #7291
Comments
https://github.com/ggerganov/llama.cpp/tree/master/examples/perplexity#perplexity
|
Does it work with a plain text file? |
It could be related to text formatting. When I was using my usual copy pasted prompt from Wikipedia, main.exe exited like that too. When I deleted the copy pasted text and just wrote "Hello", it worked well. |
Following up here, the solution was to convert the arc-challenge binary file to .json. Was able to get the baseline KLD and the resulting divergence.dat file was populated as expected. Should be good to close here, appreciate everyone's input. |
Hi there, creating an issue for a possible bug encountered while running some tests with perplexity.exe on Llama 3 earlier today. The program is exiting without running the KL-divergence calculation.
Steps to Reproduce:
Using the 8b base model, downloaded from https://huggingface.co/meta-llama/Meta-Llama-3-8B, I ran convert-hf-to-gguf.py with the following command:
So far so good, but then when I try to run perplexity.exe, it exits without outputing anything:
Here's a screenshot right before it exits, tokenizing the input appears to the final step reached:
Is there something I'm missing here? I was able to run this same command for a model based on Mistral 7b without any problems as shown in the thread linked above. Also I'm not getting any pre-tokenizer warnings when loading the fp16 GGUF for what it's worth.
Quick edit: Verified that this seems to be happening with the f32 and Q8_0 as well.
The text was updated successfully, but these errors were encountered: