diff --git a/src/langtrace_python_sdk/instrumentation/cohere/patch.py b/src/langtrace_python_sdk/instrumentation/cohere/patch.py index 7704f138..13132fb7 100644 --- a/src/langtrace_python_sdk/instrumentation/cohere/patch.py +++ b/src/langtrace_python_sdk/instrumentation/cohere/patch.py @@ -23,7 +23,6 @@ get_llm_url, get_span_name, set_event_completion, - set_event_completion_chunk, set_usage_attributes, ) from langtrace.trace_attributes import Event, LLMSpanAttributes diff --git a/src/langtrace_python_sdk/instrumentation/gemini/patch.py b/src/langtrace_python_sdk/instrumentation/gemini/patch.py index 7e7cc68d..bd585c5a 100644 --- a/src/langtrace_python_sdk/instrumentation/gemini/patch.py +++ b/src/langtrace_python_sdk/instrumentation/gemini/patch.py @@ -13,7 +13,6 @@ get_span_name, is_streaming, set_event_completion, - set_event_completion_chunk, set_span_attributes, set_usage_attributes, ) @@ -156,7 +155,6 @@ def build_streaming_response(span, response): item_to_yield = item complete_response += str(item.text) yield item_to_yield - set_event_completion_chunk(span, item.text) if hasattr(item, "usage_metadata"): usage = item.usage_metadata input_tokens = usage.prompt_token_count @@ -176,7 +174,6 @@ async def abuild_streaming_response(span, response): item_to_yield = item complete_response += str(item.text) yield item_to_yield - set_event_completion_chunk(span, item.text) if hasattr(item, "usage_metadata"): usage = item.usage_metadata input_tokens = usage.prompt_token_count diff --git a/src/langtrace_python_sdk/instrumentation/groq/patch.py b/src/langtrace_python_sdk/instrumentation/groq/patch.py index eb64eb1a..6d5ab736 100644 --- a/src/langtrace_python_sdk/instrumentation/groq/patch.py +++ b/src/langtrace_python_sdk/instrumentation/groq/patch.py @@ -31,7 +31,6 @@ get_langtrace_attributes, get_span_name, set_event_completion, - set_event_completion_chunk, set_usage_attributes, ) from langtrace_python_sdk.constants.instrumentation.common import ( @@ -245,14 +244,6 @@ def handle_streaming_response( else: content = [] - set_event_completion_chunk( - span, - ( - "".join(content) - if len(content) > 0 and content[0] is not None - else "" - ), - ) result_content.append(content[0] if len(content) > 0 else "") yield chunk finally: diff --git a/src/langtrace_python_sdk/instrumentation/ollama/patch.py b/src/langtrace_python_sdk/instrumentation/ollama/patch.py index 45551bbf..5880abf0 100644 --- a/src/langtrace_python_sdk/instrumentation/ollama/patch.py +++ b/src/langtrace_python_sdk/instrumentation/ollama/patch.py @@ -8,7 +8,6 @@ get_llm_url, get_span_name, set_event_completion, - set_event_completion_chunk, ) from langtrace_python_sdk.utils.silently_fail import silently_fail from langtrace_python_sdk.constants.instrumentation.common import SERVICE_PROVIDERS diff --git a/src/langtrace_python_sdk/version.py b/src/langtrace_python_sdk/version.py index 6d526cc8..0ec7f4c0 100644 --- a/src/langtrace_python_sdk/version.py +++ b/src/langtrace_python_sdk/version.py @@ -1 +1 @@ -__version__ = "2.2.23" +__version__ = "2.2.24"