Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

bug: RAG loses context on model response regeneration with Groq models via API key #1105

Closed
2 tasks done
silentoplayz opened this issue Mar 8, 2024 · 0 comments · Fixed by #1107
Closed
2 tasks done

Comments

@silentoplayz
Copy link
Collaborator

silentoplayz commented Mar 8, 2024

Feel free to adjust the title of this bug report, I know it's a terrible one.

Bug Report

Description

Bug Summary: The RAG feature loses context on model response regeneration when using Groq models via the API integration into Open WebUI.

Expected Behavior:
The RAG feature should maintain context and provide consistent responses throughout the conversation, even on model response regeneration.

Actual Behavior:
The RAG feature does not maintain context and provides inconsistent responses throughout the conversation when using Groq models via the API integration into Open WebUI. However, when using a locally installed LLM, the RAG feature is able to maintain context and provide consistent responses throughout the conversation.

Environment

Operating System:
Windows 11 Pro Insider Preview
Version 22H2
Installed on ‎2/‎4/‎2024
OS build 23620.1000
Experience Windows Feature Experience Pack 1000.23620.1000.0

Browser: FireFox v123.0.1 (64-bit)

Reproduction Details

Steps to Reproduce:

  1. Integrate Groq models via the API into Open WebUI
  2. Interact with the RAG feature
  3. Observe the initial response
  4. Regenerate the response
  5. Observe the loss of context in the regenerated response

Confirmation:

  • I have read and followed all the instructions provided in the README.md.
  • I am on the latest version of both Open WebUI and Ollama.

Additional Information

Note: This issue does not occur when using a locally installed LLM. The RAG feature is able to maintain context and provide consistent responses throughout the conversation when using a locally installed LLM. As far as I am aware, this issue is specific to the integration of Groq models via the API in Open WebUI, but it could potentially effect multiple APIs, being any that I don't have access to.

Edit: If anyone could replicate this issue and verify that it is indeed a bug, please share down below in the comments.

@tjbck tjbck linked a pull request Mar 9, 2024 that will close this issue
@tjbck tjbck mentioned this issue Mar 9, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

1 participant