-
Could not find a GGUF version of https://huggingface.co/intfloat/e5-mistral-7b-instruct I am trying to convert it to GGUF using convert.py but getting a KeyError: 'embed_tokens.weight' Any suggestions? Thanks,
|
Beta Was this translation helpful? Give feedback.
Replies: 4 comments 4 replies
-
Is this error because embedding models are not supported by the conversion tool? |
Beta Was this translation helpful? Give feedback.
-
were you able to get this working? |
Beta Was this translation helpful? Give feedback.
-
No. I did not know how to do it.
…On Mon, Jan 15, 2024 at 8:14 PM Greg Tanaka ***@***.***> wrote:
were you able to get this working?
—
Reply to this email directly, view it on GitHub
<#4786 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AABY4ODPOLJNMJVNABGXKITYOX5DFAVCNFSM6AAAAABBOYKJMCVHI2DSMVQWIX3LMV43SRDJONRXK43TNFXW4Q3PNVWWK3TUHM4DCMZZGY2TK>
.
You are receiving this because you authored the thread.Message ID:
***@***.***>
|
Beta Was this translation helpful? Give feedback.
-
I probably did it, just by modify the tensor_mapping.py and update TensorNameMap dictionary using names stricly from Lora adapter. Then: convert to fp16.bin base model and Lora https://huggingface.co/s3nh/intfloat-e5-mistral-7b-instruct-GGUF |
Beta Was this translation helpful? Give feedback.
I probably did it, just by modify the tensor_mapping.py and update TensorNameMap dictionary using names stricly from Lora adapter.
Upload modified one here:
https://gist.github.com/s3nh/a06f827bc492eb4b667db09d44b922e7
Then:
convert to fp16.bin base model and Lora
merge them
quantize with llama.cpp/quantize
I got feedback that It looks ok so you can give it a try and prove me wrong eventually.
https://huggingface.co/s3nh/intfloat-e5-mistral-7b-instruct-GGUF