-
-
Notifications
You must be signed in to change notification settings - Fork 562
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Feature Request] Conversational Search (RAG) with a local LLM #1732
Comments
It takes only care of the R part of RAG but yes, custom models and using GPU are supported. Check out all the subtopics of this part of the documentation: |
@piccaso Typesense does support the "AG" part of RAG, by integrating with ChatGPT / Cloduflare APIs: https://typesense.org/docs/26.0/api/conversational-search-rag.html @elliot-sawyer We don't yet have a way to integrate with local LLMs. But I'll leave this open as a feature request. |
May I know which local LLMs you're looking for? |
I don't have a particular one in mind yet - would any of the Typesense models on HuggingFace be appropriate? I'll have an NVIDIA A100 available in a couple of months to do some Typesense work with, but only on the stipulation that I use a locally downloaded LLM (no network or API keys). |
I misspoke earlier. Turns out that we actually added support for vLLM through which you can run several local LLMs. Just haven't documented it yet. Will post a link here once we update the docs. |
@jasonbosco It's nice to know that this is being done. I already started testing this myself, without documentation (heh) and ran into an authorization problem. Now |
CC: @ozanarmagan |
Thats huge. If you manage to make this easily accessible it could be quite the hype. |
Description
Is it possible to use Conversational Search (RAG) with a local LLM? The documentation suggests it is only possible with OpenAI and Cloudflare. I was wondering if any of the HuggingFace models could be used with an available GPU instead to avoid making slow network calls.
Metadata
Typesense Version: 26
OS: Linux
The text was updated successfully, but these errors were encountered: