-
Notifications
You must be signed in to change notification settings - Fork 578
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Plug & Play download & configuration of HuggingFace sentence transformer models #26696
Comments
It's a bit technical, plus people want to quantize, not quantize, and so forth. |
Closing inactive issue. |
I think we want to stick with a curated set with a guaranteed lifecycle: https://cloud.vespa.ai/en/model-hub |
Reopened! |
Context @eostis ? |
I think it would be useful to put some models on the shelves. The onnx generation is a real headache. Maybe by making Vespa Hub downloadable? |
It can be tricky: |
Sure, I get that, but it's a general problem with the python world of dependencies that I don't think we can fully address :) |
(As usual, I'll give the point of view of small projects like WooCommerce ...)
The current way of using HF models is quite technical: build scripts to download models, convert to ONNX, deploy
(https://github.com/vespa-engine/sample-apps/blob/master/transformers/src/python/setup-model.py)
I suggest to ease the process, to be able to use any HF sentence transformer model, by configuration rather than scripts.
Some possibilities:
The text was updated successfully, but these errors were encountered: