Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Plug & Play download & configuration of HuggingFace sentence transformer models #26696

Open
eostis opened this issue Apr 6, 2023 · 9 comments
Milestone

Comments

@eostis
Copy link

eostis commented Apr 6, 2023

(As usual, I'll give the point of view of small projects like WooCommerce ...)

The current way of using HF models is quite technical: build scripts to download models, convert to ONNX, deploy
(https://github.com/vespa-engine/sample-apps/blob/master/transformers/src/python/setup-model.py)

I suggest to ease the process, to be able to use any HF sentence transformer model, by configuration rather than scripts.

Some possibilities:

  • Add parameters to the docker compose file
  • Add parameters to the schema (most flexible I think)
  • Add parameters to the services.xml file
@jobergum
Copy link
Member

It's a bit technical, plus people want to quantize, not quantize, and so forth.

@johans1 johans1 added this to the later milestone Apr 12, 2023
@eostis
Copy link
Author

eostis commented Sep 12, 2023

Closing inactive issue.

@eostis eostis closed this as completed Sep 12, 2023
@bratseth
Copy link
Member

I think we want to stick with a curated set with a guaranteed lifecycle: https://cloud.vespa.ai/en/model-hub

@eostis
Copy link
Author

eostis commented Sep 28, 2023

Reopened!

@eostis eostis reopened this Sep 28, 2023
@jobergum
Copy link
Member

Context @eostis ?

@eostis
Copy link
Author

eostis commented Sep 28, 2023

I think it would be useful to put some models on the shelves. The onnx generation is a real headache. Maybe by making Vespa Hub downloadable?

@jobergum
Copy link
Member

We can probably put a few onnx models on HF. Do you think it is still painful to use Optimum to export models?

image

@jobergum
Copy link
Member

Sure, I get that, but it's a general problem with the python world of dependencies that I don't think we can fully address :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

4 participants