-
Notifications
You must be signed in to change notification settings - Fork 560
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Chunks too Large #410
Comments
This is confirmed to exist on the Main Branch as of 12/19. Reach out for the file. |
On 12/19 Build Main. It does work with the BAAI embeddings. |
What would be better.pdf |
@Patrick-Davis-MSFT , as you can see George and the team have begun working on this issue. The fix is a bit too deep and complex to make it into the v1.0 release this close to release date. So we have begun the work, but are targeting this for a hotfix after the v1.0 release. Just wanted to give you an update on the plan. |
@Patrick-Davis-MSFT , This hotfix has been applied to main in #478 . Closing this issue. |
Occurs in Delta Release
Describe the bug
When uploading some documents, we receive the error using the Azure Embeddings model ADA.
openai.error.InvalidRequestError: This model's maximum context length is 8191 tokens, however you requested 22089 tokens (22089 in your prompt; 0 for the completion). Please reduce your prompt; or completion length.
The embeddings model has a capacity of 352K TPM.
To Reproduce
Steps to reproduce the behavior:
Expected behavior
The embeddings should resolve as normal and if the chunk is too big split the chunk.
Screenshots
full error is below from 2023-12-19 18:52:43
Desktop (please complete the following information):
Alpha version details
Additional context
A commit was pushed while writing this ticket for the auto scaler. I will try with the latest and update once done.
The text was updated successfully, but these errors were encountered: