New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Problem loading metadata of gguf file #2152
Comments
You may want to use the latest github version as 0.4.1 may well not be compatible with phi-3. |
Seems that there was a "silent" change of the naming convention in phi-3 gguf models, see #2154 , candle now supports both the old and the new naming convention in the quantized-phi example, |
I think the real reason caused the problem is this.
However, it appears that phi3 can only be converted using convert-hf-to-gguf.py due to encountering an NotImplementedError with the message: Unknown rope scaling type: su. This inconsistency in conversion methods seems to have led to the problem. The left model in the screenshot was converted using I am wondering if candle could auto detect the architecture from the gguf model converted by |
I would have thought that we support both methods now, the |
I encountered an error while executing the example quantized-phi, which I slightly modified. However, I suspect the issue might not be with my modifications.
The problem seems to be related to the function
candle_transformers::models::quantized_llama::ModelWeights::from_gguf
. It appears to be unable to locate the necessary metadata from the model. This is interesting because Hugging Face is able to display the model's metadata correctly.Here are some screenshots for further reference:
I would appreciate any assistance in resolving this issue. Thank you in advance.
Full Code
The text was updated successfully, but these errors were encountered: