From 604ea317d62799b658f9ea6fa76c8228f4f8e1bb Mon Sep 17 00:00:00 2001 From: Jinglei Ren Date: Fri, 12 May 2023 11:07:22 +0800 Subject: [PATCH] Fix bug of overwriting request token number --- .gitignore | 1 + .vscode/settings.json | 8 -------- devchat/openai/openai_prompt.py | 2 +- 3 files changed, 2 insertions(+), 9 deletions(-) delete mode 100644 .vscode/settings.json diff --git a/.gitignore b/.gitignore index 8e6d86ee..62c734cc 100644 --- a/.gitignore +++ b/.gitignore @@ -1,4 +1,5 @@ .DS_Store +.vscode # PyCharm .idea/ diff --git a/.vscode/settings.json b/.vscode/settings.json deleted file mode 100644 index c2378c84..00000000 --- a/.vscode/settings.json +++ /dev/null @@ -1,8 +0,0 @@ -{ - "python.linting.enabled": true, - "python.linting.pylintEnabled": true, - "python.linting.flake8Enabled": true, - "python.linting.flake8Args": [ - "--max-line-length=100" - ], -} \ No newline at end of file diff --git a/devchat/openai/openai_prompt.py b/devchat/openai/openai_prompt.py index 58ff72c9..d27cd0c8 100644 --- a/devchat/openai/openai_prompt.py +++ b/devchat/openai/openai_prompt.py @@ -77,7 +77,7 @@ def set_request(self, content: str) -> int: raise ValueError("The request cannot be empty.") message = OpenAIMessage(content, 'user') self._new_messages['request'] = message - self._request_tokens = message_tokens(message.to_dict(), self.model) + self._request_tokens += message_tokens(message.to_dict(), self.model) def set_response(self, response_str: str): """