-
Notifications
You must be signed in to change notification settings - Fork 269
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Feature/add local llama tut #278
Conversation
The latest updates on your projects. Learn more about Vercel for Git ↗︎
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
❌ Changes requested.
- Reviewed the entire pull request up to b0551ad
- Looked at
1032
lines of code in26
files - Took 1 minute and 30 seconds to review
More info
- Skipped
0
files when reviewing. - Skipped posting
0
additional comments because they didn't meet confidence threshold of50%
.
Workflow ID: wflow_oP3SSoC2WrK7CgfW
Want Ellipsis to fix these issues? Tag @ellipsis-dev
in a comment. We'll respond in a few minutes. Learn more here.
@@ -51,21 +49,22 @@ def search(self, query): | |||
print(body[:500]) | |||
print("\n") | |||
|
|||
def rag_completion(self, query): | |||
def rag_completion(self, query, model="gpt-4-turbo-preview"): |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The model parameter is hardcoded to 'gpt-4-turbo-preview'. Consider making this a configurable parameter to allow users to specify the model they want to use.
def rag_completion(self, query, model="gpt-4-turbo-preview"): | |
def rag_completion(self, query, model): |
Summary:
This PR adds support for local LLM providers
ollama
andLlama.cpp
in the R2R framework, along with updates to the documentation, example code, and configuration files to guide users on how to use this new feature.Key points:
ollama
andLlama.cpp
in/r2r/llms/__init__.py
and/r2r/llms/llama_cpp/base.py
.config.json
examples to include configurations for local LLM providers in/r2r/examples/configs/local_ollama.json
and/r2r/examples/configs/local_llama_cpp.json
./r2r/examples/servers/basic_pipeline.py
to allow selection of configuration based on the desired LLM provider./r2r/examples/clients/run_basic_client.py
to demonstrate how to use the new local LLM providers./docs/pages/getting-started/basic-example.mdx
,/docs/pages/getting-started/configure-your-pipeline.mdx
,/docs/pages/providers/evals.mdx
, and/docs/pages/providers/llms.mdx
to reflect the new feature./docs/pages/tutorials/local_rag.mdx
on how to run a local RAG pipeline with R2R.Generated with ❤️ by ellipsis.dev