-
-
Notifications
You must be signed in to change notification settings - Fork 1.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feature: embedding support #70
Comments
Just curious to find out what is the use/purpose of embeddings above. For the following use case of Retrieval Augmented Data QA: Can't we use the following embedding models? I plan to use
Please advise. Thank you. |
embeddings support has been merged to master. It is experimental and currently it's available only on To enable it you can set |
I've published a sample using embeddings over here: https://github.com/go-skynet/LocalAI/tree/master/examples/query_data |
further optimizations in #222 - now embeddings can be used with bert on any model - and there is also a huge performance impact! |
Hello! I am trying to run a gpt4all-j model for building a local chatbot. How can I use an embedding using BERT and implement it for chat completions endpoint? Currently, I am running it on Mac Mini i7, 32gb RAM. I am planning to upgrade it to a higher resource(vRAM) cloud server in future. Is it possible to make a fast chatbot API using own document embeddings? |
Thank you for the example! But it can't be included in the API? Currently I think you run those commands inside the container, right? Is there already the scenario that calling a certain path executes the query on the documents? |
Add support to embeddings to the API and the llama backend: https://github.com/ggerganov/llama.cpp/blob/e4422e299c10c7e84c8e987770ef40d31905a76b/llama.cpp#L2160
The text was updated successfully, but these errors were encountered: