Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion aider/__init__.py
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
from packaging import version

__version__ = "0.88.21.dev"
__version__ = "0.88.22.dev"
safe_version = __version__

try:
Expand Down
10 changes: 5 additions & 5 deletions aider/coders/base_coder.py
Original file line number Diff line number Diff line change
Expand Up @@ -2308,7 +2308,7 @@ def _print_tool_call_info(self, server_tool_calls):
# Parse and format arguments as headers with values
if tool_call.function.arguments:
# Only do JSON unwrapping for tools containing "replace" in their name
if (
if tool_call.get("function", {}).get("name") is not None and (
"replace" in tool_call.function.name.lower()
or "insert" in tool_call.function.name.lower()
or "update" in tool_call.function.name.lower()
Expand Down Expand Up @@ -2926,6 +2926,7 @@ def show_send_output(self, completion):
async def show_send_output_stream(self, completion):
received_content = False
id_index_dict = dict()
self._last_known_tool_index = 0

async for chunk in completion:
# Check if confirmation is in progress and wait if needed
Expand Down Expand Up @@ -2959,14 +2960,13 @@ async def show_send_output_stream(self, completion):
len(self.partial_response_tool_calls) - 1
)
elif tool_call_chunk.id is None:
if len(self.partial_response_tool_calls) <= index:
self.partial_response_tool_calls.extend(
[{}] * (index - len(self.partial_response_tool_calls) + 1)
)
index = self._last_known_tool_index

if tool_call_chunk.id is not None:
index = id_index_dict[tool_call_chunk.id]

self._last_known_tool_index = index

if tool_call_chunk.id:
self.partial_response_tool_calls[index]["id"] = tool_call_chunk.id
if tool_call_chunk.type:
Expand Down
66 changes: 39 additions & 27 deletions aider/resources/model-metadata.json
Original file line number Diff line number Diff line change
Expand Up @@ -6523,14 +6523,14 @@
"supports_vision": true
},
"deepseek-chat": {
"cache_read_input_token_cost": 6e-08,
"input_cost_per_token": 6e-07,
"cache_read_input_token_cost": 3e-08,
"input_cost_per_token": 3e-07,
"litellm_provider": "deepseek",
"max_input_tokens": 131072,
"max_input_tokens": 128000,
"max_output_tokens": 8192,
"max_tokens": 131072,
"max_tokens": 128000,
"mode": "chat",
"output_cost_per_token": 1.7e-06,
"output_cost_per_token": 5e-07,
"source": "https://api-docs.deepseek.com/quick_start/pricing",
"supported_endpoints": [
"/v1/chat/completions"
Expand All @@ -6544,14 +6544,14 @@
"supports_tool_choice": true
},
"deepseek-reasoner": {
"cache_read_input_token_cost": 6e-08,
"input_cost_per_token": 6e-07,
"cache_read_input_token_cost": 3e-08,
"input_cost_per_token": 3e-07,
"litellm_provider": "deepseek",
"max_input_tokens": 131072,
"max_input_tokens": 128000,
"max_output_tokens": 65536,
"max_tokens": 131072,
"max_tokens": 128000,
"mode": "chat",
"output_cost_per_token": 1.7e-06,
"output_cost_per_token": 5e-07,
"source": "https://api-docs.deepseek.com/quick_start/pricing",
"supported_endpoints": [
"/v1/chat/completions"
Expand Down Expand Up @@ -8467,19 +8467,24 @@
"supports_tool_choice": true
},
"deepseek/deepseek-chat": {
"cache_creation_input_token_cost": 0.0,
"cache_read_input_token_cost": 7e-08,
"input_cost_per_token": 2.7e-07,
"input_cost_per_token_cache_hit": 7e-08,
"cache_read_input_token_cost": 3e-08,
"input_cost_per_token": 3e-07,
"litellm_provider": "deepseek",
"max_input_tokens": 65536,
"max_input_tokens": 128000,
"max_output_tokens": 8192,
"max_tokens": 8192,
"max_tokens": 128000,
"mode": "chat",
"output_cost_per_token": 1.1e-06,
"supports_assistant_prefill": true,
"output_cost_per_token": 5e-07,
"source": "https://api-docs.deepseek.com/quick_start/pricing",
"supported_endpoints": [
"/v1/chat/completions"
],
"supports_function_calling": true,
"supports_native_streaming": true,
"supports_parallel_function_calling": true,
"supports_prompt_caching": true,
"supports_response_schema": true,
"supports_system_messages": true,
"supports_tool_choice": true
},
"deepseek/deepseek-coder": {
Expand Down Expand Up @@ -8512,19 +8517,26 @@
"supports_tool_choice": true
},
"deepseek/deepseek-reasoner": {
"input_cost_per_token": 5.5e-07,
"input_cost_per_token_cache_hit": 1.4e-07,
"cache_read_input_token_cost": 3e-08,
"input_cost_per_token": 3e-07,
"litellm_provider": "deepseek",
"max_input_tokens": 65536,
"max_output_tokens": 8192,
"max_tokens": 8192,
"max_input_tokens": 128000,
"max_output_tokens": 65536,
"max_tokens": 128000,
"mode": "chat",
"output_cost_per_token": 2.19e-06,
"supports_assistant_prefill": true,
"supports_function_calling": true,
"output_cost_per_token": 5e-07,
"source": "https://api-docs.deepseek.com/quick_start/pricing",
"supported_endpoints": [
"/v1/chat/completions"
],
"supports_function_calling": false,
"supports_native_streaming": true,
"supports_parallel_function_calling": false,
"supports_prompt_caching": true,
"supports_reasoning": true,
"supports_tool_choice": true
"supports_response_schema": true,
"supports_system_messages": true,
"supports_tool_choice": false
},
"deepseek/deepseek-v3": {
"cache_creation_input_token_cost": 0.0,
Expand Down
Loading