Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

bug: RAG follow-up questions show lost in the middle issue #2519

Closed
Tracked by #1595
Propheticus opened this issue Mar 27, 2024 · 3 comments
Closed
Tracked by #1595

bug: RAG follow-up questions show lost in the middle issue #2519

Propheticus opened this issue Mar 27, 2024 · 3 comments
Assignees
Labels
type: bug Something isn't working

Comments

@Propheticus
Copy link

Propheticus commented Mar 27, 2024

Describe the bug
After attaching even a small 8 page document; When asking a follow-up questions about a topic that is discussed in the middle, no correct answers are given.
When you ask a question about this topic in the first question of a thread, the info is found and the answer is correct.

Steps to reproduce
Steps to reproduce the behaviour:

  1. Attach a document in the first chat of a thread and ask for a summary
  2. In the second prompt ask a follow-up question about a topic discussed midway the doc.
  3. See the response stating nothing is said about that topic in the provided context. (false)

Expected behaviour
Using a model like Mistral 7B v0.2 I expect no lost in the middle issue, even when using a larger context.
In comparison: when I use AnythingLLM for embedding/vectorDB, either with or without pinning (pasting of the entire doc in context), the questions áre properly answered.

Screenshots
If applicable, add screenshots to help explain your issue.
Second question, nothing found:
image
First question, found without issue:
image
Second question via AnythingLLM, found without issue:
image

Environment details

  • Operating System: Win 11 Pro N x64
  • Jan Version: 0.4.9-345
  • Processor: Ryzen 7 7700X
  • RAM: 32GB
  • GPU: AMD RX 6800XT 16GB
  • Model: Mistral instruct 7B v0.2 Q5_K_M, max response tokens: 4096, Context length: 30K tokens.
  • Retrieval settings:
    image

Logs
No errors. They state a successful response was generated....which it was, just the content is of low quality.

@Propheticus Propheticus added the type: bug Something isn't working label Mar 27, 2024
@louis-jan louis-jan added this to the v0.4.11 milestone Apr 3, 2024
@Van-QA Van-QA modified the milestones: v0.4.11, v0.4.12 Apr 9, 2024
@Van-QA Van-QA modified the milestones: v0.4.12, v0.4.13 Apr 17, 2024
@haliliceylan
Copy link

+1, Same problem here.

@louis-jan
Copy link
Contributor

We will rework on RAG with separate embedding models.

@louis-jan louis-jan modified the milestones: v0.4.13, v0.4.14 Apr 26, 2024
@Van-QA Van-QA modified the milestones: v.0.5.1 🍖 Kebap, v0.6.0 📜 Jun 6, 2024
@0xSage
Copy link
Contributor

0xSage commented Jun 11, 2024

dupe of janhq/cortex.cpp#1595

@0xSage 0xSage closed this as completed Jun 11, 2024
@imtuyethan imtuyethan removed this from the v0.X.0 📜 milestone Sep 11, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
type: bug Something isn't working
Projects
Archived in project
Development

No branches or pull requests

6 participants