-
Notifications
You must be signed in to change notification settings - Fork 5.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Embedding results have changed in v0.1.32 #3777
Comments
I also have this issue. Also (even in 0.1.30), it never loads the full
@SunmeetOberoi can you please also cross check Ollama logs using |
@jmorganca Looks like 0.1.32 is using the wrong model config parameter to determine max context length. Details [here] (#3727 (comment)) Also see Readme.md in the nomic GGUF file repository for this model. |
Hi @Kanishk-Kumar, I tried it out and yes I am also seeing that log message and the n_ctx as 2048
Also, the embeddings are not the same amongst these versions. As @jimscard correctly pointed out the nomic GGUF readme file does mentions something related to this as well which might help.
Since this context length issue is existing in v0.1.31 as well and is a model specific issue. I think its not related to the different embedding values problem mentioned here and can be tracked separately in #3727 |
Guys, any luck here? I just bumped to v0.1.33 and this issue still exists. |
I've pinpointed the issue to this commit 5ec12ce |
What is the issue?
The values for embedding have changed in 0.1.32 release.
I used an older version of ollama to complete a POC for categorization of some data it all went fine. Now when I was trying to implement the solution I could see the search result were way off. Almost all categories had a 50-60% similarity with every input value. After trying to fix my script for hours I thought to downgrade ollama and that worked.
This is only happening in 0.1.32, I tested the same code on 0.1.29, 0.1.30 and 0.1.31 it is working consistently and accurately.
Attaching a sample python script to test out the observation:
Results:
Output for v0.1.29
Output for v0.1.30
Output for v0.1.31
Output for v0.1.32
Python libraries version
OS
Windows, WSL2
GPU
Nvidia
CPU
Intel
Ollama version
0.1.32
The text was updated successfully, but these errors were encountered: