Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[BUG] Slow retrieval of documents #305

Closed
jeremylatorre opened this issue May 13, 2024 · 9 comments
Closed

[BUG] Slow retrieval of documents #305

jeremylatorre opened this issue May 13, 2024 · 9 comments
Labels
needs-info Need more information such as reproduction step stale Issue labeled as stale will be closed automatically

Comments

@jeremylatorre
Copy link
Contributor

Describe the bug

When using RAG, I noticed that performances where very low when trying to retrieve information from documents.
Any lead about what is causing this. Clearly the difference between rag and standard bot is huge at that time.
It might also depend on number of documents.

@statefb statefb added the needs-info Need more information such as reproduction step label May 14, 2024
@statefb
Copy link
Contributor

statefb commented May 14, 2024

Could you provide the detail? Please do not ignore the issue template.

@jeremylatorre
Copy link
Contributor Author

When I use my bot, it takes sometimes 17s before starting to render the content.

The first step [Retrieve Knowledge] looks ok, but after that, cursor stay at the beginning of the line for a very long time before starting to write the response.

This doesn't happen when using the LLM without documents, and it seems to increase with the number of document in the knowledge base.

@statefb
Copy link
Contributor

statefb commented May 14, 2024

Could you check the items in postgres? We can use management console to query and the table definition can be reffered here.

@jeremylatorre
Copy link
Contributor Author

Could it be related to the instantion of a Lambda function? The behavior only appear on the first inference for a new conversation

@statefb
Copy link
Contributor

statefb commented May 16, 2024

Have you forked and customized this sample? Large container causes longer cold start. You can check whether the invocation is cold or not by reffering cloudwatch logs.
https://stackoverflow.com/questions/47061146/how-can-i-detect-cold-starts-in-aws-lambda

@jeremylatorre
Copy link
Contributor Author

Ok I have digged a llittle bit on the logs and I found that the POST call to related-documents is very long by itself:

Request POST /conversation/related-documents (15:15:32 - 15:15:44)
Duration : 11697 ms

I'm not currently up to date with the codebase and we will perform a full update soon. I'll keep you updated on this.

@statefb
Copy link
Contributor

statefb commented Jun 13, 2024

@jeremylatorre Thank you for the update. The query part to postgres seems to making the very long latency. For the cost effectiveness, the ACU configured at minimum value by default. Choosing larger value might solve the problem .

Copy link
Contributor

This issue has been labeled as "stale" due to no response by the reporter within 1 month (and 14 days after last commented by someone). And it will be closed automatically 14 days later if not responded.

@github-actions github-actions bot added the stale Issue labeled as stale will be closed automatically label Jul 12, 2024
Copy link
Contributor

This issue has been closed due to no response within 14 days after labeled as "stale", 14 days after last reopened, and 14 days after last commented.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
needs-info Need more information such as reproduction step stale Issue labeled as stale will be closed automatically
Projects
None yet
Development

No branches or pull requests

2 participants