-
Notifications
You must be signed in to change notification settings - Fork 105
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
fix: fix model sizes in supported models lists #167
Conversation
I went through list of supported models and found some discrepancies. There were various cases, I'll try to describe how I chose the sizes:
intfloat/multilingual-e5-large had a wrong link, it was named differently in the cloud, I renamed it there. |
Check out this pull request on See visual diffs & provide feedback on Jupyter Notebooks. Powered by ReviewNB |
Wherever we've quantized models which meet the allclose tests — we'll use that as our default, without supporting the direct onnx import to avoid confusion. We also prefer HF over URL download. The split here would've changed the underlying default model without informing the user, so I've tried to make that consistent for all models here |
@@ -70,15 +61,6 @@ | |||
"dim": 384, | |||
"description": "Fast and Default English model", | |||
"size_in_GB": 0.13, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I believe this should be adjusted since the quantized model is 2 times smaller
No description provided.