You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
During our routine tests, we've been noticing that at least one question fails in each test run, reducing the overall average. The issue seems to be occurring randomly and is not reproducible in local environments, leading to the suspicion that it may be related to issues on deployment or OpenAI's side.
Details:
From analysis of the celeryworker.log and the information available in the admin panel, it seems that the failures might be due to hitting the rate limit set by OpenAI. The error message reads: "Rate limit reached for 10KTPM-200RPM in organization org-xxxxx tokens per min. Limit: 10000 / min"
Currently, langchain already has a retry mechanism in place, where it retries a failed request in 2^x seconds intervals if it's a rate limit issue. However, this doesn't seem to be sufficient to handle the current situation.
Proposed Solution:
Considering we don't need test results urgently, we could increase the maximum wait time for retries, say up to 3 minutes. This should provide a larger buffer to avoid hitting the rate limit and might potentially resolve the issue.
Next Steps:
Investigate if the current langchain configuration allows to modify the maximum wait time for retries.
If possible, adjust the maximum wait time to 3 minutes and monitor if this reduces or eliminates the random test failures.
The text was updated successfully, but these errors were encountered:
During our routine tests, we've been noticing that at least one question fails in each test run, reducing the overall average. The issue seems to be occurring randomly and is not reproducible in local environments, leading to the suspicion that it may be related to issues on deployment or OpenAI's side.
Details:
From analysis of the
celeryworker.log
and the information available in the admin panel, it seems that the failures might be due to hitting the rate limit set by OpenAI. The error message reads: "Rate limit reached for 10KTPM-200RPM in organization org-xxxxx tokens per min. Limit: 10000 / min"Currently, langchain already has a retry mechanism in place, where it retries a failed request in 2^x seconds intervals if it's a rate limit issue. However, this doesn't seem to be sufficient to handle the current situation.
Proposed Solution:
Considering we don't need test results urgently, we could increase the maximum wait time for retries, say up to 3 minutes. This should provide a larger buffer to avoid hitting the rate limit and might potentially resolve the issue.
Next Steps:
The text was updated successfully, but these errors were encountered: