diff --git a/.stats.yml b/.stats.yml index 71a56c4c..037cba0b 100644 --- a/.stats.yml +++ b/.stats.yml @@ -1,4 +1,4 @@ configured_endpoints: 95 -openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/openai%2Fopenai-32de3bc513663c5fac922c49be41c222b6ee8c0b841d8966bcdfa489d441daa3.yml +openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/openai%2Fopenai-44b20fa9d24544217fe6bb48852037537030a1ad29b202936425110744fe66fb.yml openapi_spec_hash: ea86343b5e9858a74e85da8ab2c532f6 -config_hash: 43dc8df20ffec9d1503f91866cb2b7d9 +config_hash: 69e3afd56ccb0f0f822a7a9dc130fc99 diff --git a/README.md b/README.md index 7a859626..85e4c03d 100644 --- a/README.md +++ b/README.md @@ -73,7 +73,7 @@ We provide support for streaming responses using Server Side Events (SSE). ```ruby stream = - openai.chat.completions.create_streaming( + openai.chat.completions.stream_raw( messages: [{ role: "user", content: "Say this is a test" diff --git a/lib/openai/models/beta/thread_create_and_run_params.rb b/lib/openai/models/beta/thread_create_and_run_params.rb index f2349d0f..4612ebd8 100644 --- a/lib/openai/models/beta/thread_create_and_run_params.rb +++ b/lib/openai/models/beta/thread_create_and_run_params.rb @@ -5,7 +5,7 @@ module Models module Beta # @see OpenAI::Resources::Beta::Threads#create_and_run # - # @see OpenAI::Resources::Beta::Threads#create_and_run_streaming + # @see OpenAI::Resources::Beta::Threads#stream_raw class ThreadCreateAndRunParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter diff --git a/lib/openai/models/beta/threads/run.rb b/lib/openai/models/beta/threads/run.rb index 6b71de5b..c6436c0d 100644 --- a/lib/openai/models/beta/threads/run.rb +++ b/lib/openai/models/beta/threads/run.rb @@ -6,7 +6,7 @@ module Beta module Threads # @see OpenAI::Resources::Beta::Threads::Runs#create # - # @see OpenAI::Resources::Beta::Threads::Runs#create_streaming + # @see OpenAI::Resources::Beta::Threads::Runs#create_stream_raw class Run < OpenAI::Internal::Type::BaseModel # @!attribute id # The identifier, which can be referenced in API endpoints. diff --git a/lib/openai/models/beta/threads/run_create_params.rb b/lib/openai/models/beta/threads/run_create_params.rb index 3e2776ba..a058e78e 100644 --- a/lib/openai/models/beta/threads/run_create_params.rb +++ b/lib/openai/models/beta/threads/run_create_params.rb @@ -6,7 +6,7 @@ module Beta module Threads # @see OpenAI::Resources::Beta::Threads::Runs#create # - # @see OpenAI::Resources::Beta::Threads::Runs#create_streaming + # @see OpenAI::Resources::Beta::Threads::Runs#create_stream_raw class RunCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter diff --git a/lib/openai/models/beta/threads/run_submit_tool_outputs_params.rb b/lib/openai/models/beta/threads/run_submit_tool_outputs_params.rb index a3e1b7d9..2243cb1d 100644 --- a/lib/openai/models/beta/threads/run_submit_tool_outputs_params.rb +++ b/lib/openai/models/beta/threads/run_submit_tool_outputs_params.rb @@ -6,7 +6,7 @@ module Beta module Threads # @see OpenAI::Resources::Beta::Threads::Runs#submit_tool_outputs # - # @see OpenAI::Resources::Beta::Threads::Runs#submit_tool_outputs_streaming + # @see OpenAI::Resources::Beta::Threads::Runs#submit_tool_outputs_stream_raw class RunSubmitToolOutputsParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter diff --git a/lib/openai/models/chat/chat_completion.rb b/lib/openai/models/chat/chat_completion.rb index 4ea12d1b..97ee34db 100644 --- a/lib/openai/models/chat/chat_completion.rb +++ b/lib/openai/models/chat/chat_completion.rb @@ -5,7 +5,7 @@ module Models module Chat # @see OpenAI::Resources::Chat::Completions#create # - # @see OpenAI::Resources::Chat::Completions#create_streaming + # @see OpenAI::Resources::Chat::Completions#stream_raw class ChatCompletion < OpenAI::Internal::Type::BaseModel # @!attribute id # A unique identifier for the chat completion. diff --git a/lib/openai/models/chat/completion_create_params.rb b/lib/openai/models/chat/completion_create_params.rb index 4c378db0..65406125 100644 --- a/lib/openai/models/chat/completion_create_params.rb +++ b/lib/openai/models/chat/completion_create_params.rb @@ -5,7 +5,7 @@ module Models module Chat # @see OpenAI::Resources::Chat::Completions#create # - # @see OpenAI::Resources::Chat::Completions#create_streaming + # @see OpenAI::Resources::Chat::Completions#stream_raw class CompletionCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter diff --git a/lib/openai/models/responses/response.rb b/lib/openai/models/responses/response.rb index 5280f3ed..3ad5a2be 100644 --- a/lib/openai/models/responses/response.rb +++ b/lib/openai/models/responses/response.rb @@ -5,7 +5,7 @@ module Models module Responses # @see OpenAI::Resources::Responses#create # - # @see OpenAI::Resources::Responses#create_streaming + # @see OpenAI::Resources::Responses#stream_raw class Response < OpenAI::Internal::Type::BaseModel # @!attribute id # Unique identifier for this Response. diff --git a/lib/openai/models/responses/response_create_params.rb b/lib/openai/models/responses/response_create_params.rb index 07d25b0a..85b349d0 100644 --- a/lib/openai/models/responses/response_create_params.rb +++ b/lib/openai/models/responses/response_create_params.rb @@ -5,7 +5,7 @@ module Models module Responses # @see OpenAI::Resources::Responses#create # - # @see OpenAI::Resources::Responses#create_streaming + # @see OpenAI::Resources::Responses#stream_raw class ResponseCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter diff --git a/lib/openai/resources/beta/threads.rb b/lib/openai/resources/beta/threads.rb index 2902111e..b72ad867 100644 --- a/lib/openai/resources/beta/threads.rb +++ b/lib/openai/resources/beta/threads.rb @@ -94,8 +94,7 @@ def delete(thread_id, params = {}) ) end - # See {OpenAI::Resources::Beta::Threads#create_and_run_streaming} for streaming - # counterpart. + # See {OpenAI::Resources::Beta::Threads#stream_raw} for streaming counterpart. # # Create a thread and run it in one request. # @@ -124,7 +123,7 @@ def delete(thread_id, params = {}) def create_and_run(params) parsed, options = OpenAI::Models::Beta::ThreadCreateAndRunParams.dump_request(params) if parsed[:stream] - message = "Please use `#create_and_run_streaming` for the streaming use case." + message = "Please use `#stream_raw` for the streaming use case." raise ArgumentError.new(message) end @client.request( @@ -141,7 +140,7 @@ def create_and_run(params) # # Create a thread and run it in one request. # - # @overload create_and_run_streaming(assistant_id:, instructions: nil, max_completion_tokens: nil, max_prompt_tokens: nil, metadata: nil, model: nil, parallel_tool_calls: nil, response_format: nil, temperature: nil, thread: nil, tool_choice: nil, tool_resources: nil, tools: nil, top_p: nil, truncation_strategy: nil, request_options: {}) + # @overload stream_raw(assistant_id:, instructions: nil, max_completion_tokens: nil, max_prompt_tokens: nil, metadata: nil, model: nil, parallel_tool_calls: nil, response_format: nil, temperature: nil, thread: nil, tool_choice: nil, tool_resources: nil, tools: nil, top_p: nil, truncation_strategy: nil, request_options: {}) # # @param assistant_id [String] # @param instructions [String, nil] @@ -163,7 +162,7 @@ def create_and_run(params) # @return [OpenAI::Internal::Stream] # # @see OpenAI::Models::Beta::ThreadCreateAndRunParams - def create_and_run_streaming(params) + def stream_raw(params) parsed, options = OpenAI::Models::Beta::ThreadCreateAndRunParams.dump_request(params) unless parsed.fetch(:stream, true) message = "Please use `#create_and_run` for the non-streaming use case." diff --git a/lib/openai/resources/beta/threads/runs.rb b/lib/openai/resources/beta/threads/runs.rb index 61d7e284..8ebefd9c 100644 --- a/lib/openai/resources/beta/threads/runs.rb +++ b/lib/openai/resources/beta/threads/runs.rb @@ -8,7 +8,7 @@ class Runs # @return [OpenAI::Resources::Beta::Threads::Runs::Steps] attr_reader :steps - # See {OpenAI::Resources::Beta::Threads::Runs#create_streaming} for streaming + # See {OpenAI::Resources::Beta::Threads::Runs#create_stream_raw} for streaming # counterpart. # # Create a run. @@ -41,7 +41,7 @@ class Runs def create(thread_id, params) parsed, options = OpenAI::Models::Beta::Threads::RunCreateParams.dump_request(params) if parsed[:stream] - message = "Please use `#create_streaming` for the streaming use case." + message = "Please use `#create_stream_raw` for the streaming use case." raise ArgumentError.new(message) end query_params = [:include] @@ -60,7 +60,7 @@ def create(thread_id, params) # # Create a run. # - # @overload create_streaming(thread_id, assistant_id:, include: nil, additional_instructions: nil, additional_messages: nil, instructions: nil, max_completion_tokens: nil, max_prompt_tokens: nil, metadata: nil, model: nil, parallel_tool_calls: nil, reasoning_effort: nil, response_format: nil, temperature: nil, tool_choice: nil, tools: nil, top_p: nil, truncation_strategy: nil, request_options: {}) + # @overload create_stream_raw(thread_id, assistant_id:, include: nil, additional_instructions: nil, additional_messages: nil, instructions: nil, max_completion_tokens: nil, max_prompt_tokens: nil, metadata: nil, model: nil, parallel_tool_calls: nil, reasoning_effort: nil, response_format: nil, temperature: nil, tool_choice: nil, tools: nil, top_p: nil, truncation_strategy: nil, request_options: {}) # # @param thread_id [String] # @param assistant_id [String] @@ -85,7 +85,7 @@ def create(thread_id, params) # @return [OpenAI::Internal::Stream] # # @see OpenAI::Models::Beta::Threads::RunCreateParams - def create_streaming(thread_id, params) + def create_stream_raw(thread_id, params) parsed, options = OpenAI::Models::Beta::Threads::RunCreateParams.dump_request(params) unless parsed.fetch(:stream, true) message = "Please use `#create` for the non-streaming use case." @@ -208,7 +208,7 @@ def cancel(run_id, params) ) end - # See {OpenAI::Resources::Beta::Threads::Runs#submit_tool_outputs_streaming} for + # See {OpenAI::Resources::Beta::Threads::Runs#submit_tool_outputs_stream_raw} for # streaming counterpart. # # When a run has the `status: "requires_action"` and `required_action.type` is @@ -229,7 +229,7 @@ def cancel(run_id, params) def submit_tool_outputs(run_id, params) parsed, options = OpenAI::Models::Beta::Threads::RunSubmitToolOutputsParams.dump_request(params) if parsed[:stream] - message = "Please use `#submit_tool_outputs_streaming` for the streaming use case." + message = "Please use `#submit_tool_outputs_stream_raw` for the streaming use case." raise ArgumentError.new(message) end thread_id = @@ -253,7 +253,7 @@ def submit_tool_outputs(run_id, params) # tool calls once they're all completed. All outputs must be submitted in a single # request. # - # @overload submit_tool_outputs_streaming(run_id, thread_id:, tool_outputs:, request_options: {}) + # @overload submit_tool_outputs_stream_raw(run_id, thread_id:, tool_outputs:, request_options: {}) # # @param run_id [String] # @param thread_id [String] @@ -263,7 +263,7 @@ def submit_tool_outputs(run_id, params) # @return [OpenAI::Internal::Stream] # # @see OpenAI::Models::Beta::Threads::RunSubmitToolOutputsParams - def submit_tool_outputs_streaming(run_id, params) + def submit_tool_outputs_stream_raw(run_id, params) parsed, options = OpenAI::Models::Beta::Threads::RunSubmitToolOutputsParams.dump_request(params) unless parsed.fetch(:stream, true) message = "Please use `#submit_tool_outputs` for the non-streaming use case." diff --git a/lib/openai/resources/chat/completions.rb b/lib/openai/resources/chat/completions.rb index c1d28d4a..499c9f52 100644 --- a/lib/openai/resources/chat/completions.rb +++ b/lib/openai/resources/chat/completions.rb @@ -7,8 +7,7 @@ class Completions # @return [OpenAI::Resources::Chat::Completions::Messages] attr_reader :messages - # See {OpenAI::Resources::Chat::Completions#create_streaming} for streaming - # counterpart. + # See {OpenAI::Resources::Chat::Completions#stream_raw} for streaming counterpart. # # **Starting a new project?** We recommend trying # [Responses](https://platform.openai.com/docs/api-reference/responses) to take @@ -68,7 +67,7 @@ class Completions def create(params) parsed, options = OpenAI::Models::Chat::CompletionCreateParams.dump_request(params) if parsed[:stream] - message = "Please use `#create_streaming` for the streaming use case." + message = "Please use `#stream_raw` for the streaming use case." raise ArgumentError.new(message) end @client.request( @@ -100,7 +99,7 @@ def create(params) # unsupported parameters in reasoning models, # [refer to the reasoning guide](https://platform.openai.com/docs/guides/reasoning). # - # @overload create_streaming(messages:, model:, audio: nil, frequency_penalty: nil, function_call: nil, functions: nil, logit_bias: nil, logprobs: nil, max_completion_tokens: nil, max_tokens: nil, metadata: nil, modalities: nil, n: nil, parallel_tool_calls: nil, prediction: nil, presence_penalty: nil, reasoning_effort: nil, response_format: nil, seed: nil, service_tier: nil, stop: nil, store: nil, stream_options: nil, temperature: nil, tool_choice: nil, tools: nil, top_logprobs: nil, top_p: nil, user: nil, web_search_options: nil, request_options: {}) + # @overload stream_raw(messages:, model:, audio: nil, frequency_penalty: nil, function_call: nil, functions: nil, logit_bias: nil, logprobs: nil, max_completion_tokens: nil, max_tokens: nil, metadata: nil, modalities: nil, n: nil, parallel_tool_calls: nil, prediction: nil, presence_penalty: nil, reasoning_effort: nil, response_format: nil, seed: nil, service_tier: nil, stop: nil, store: nil, stream_options: nil, temperature: nil, tool_choice: nil, tools: nil, top_logprobs: nil, top_p: nil, user: nil, web_search_options: nil, request_options: {}) # # @param messages [Array] # @param model [String, Symbol, OpenAI::Models::ChatModel] @@ -137,7 +136,7 @@ def create(params) # @return [OpenAI::Internal::Stream] # # @see OpenAI::Models::Chat::CompletionCreateParams - def create_streaming(params) + def stream_raw(params) parsed, options = OpenAI::Models::Chat::CompletionCreateParams.dump_request(params) unless parsed.fetch(:stream, true) message = "Please use `#create` for the non-streaming use case." diff --git a/lib/openai/resources/responses.rb b/lib/openai/resources/responses.rb index bb369eb8..63a4abef 100644 --- a/lib/openai/resources/responses.rb +++ b/lib/openai/resources/responses.rb @@ -6,7 +6,7 @@ class Responses # @return [OpenAI::Resources::Responses::InputItems] attr_reader :input_items - # See {OpenAI::Resources::Responses#create_streaming} for streaming counterpart. + # See {OpenAI::Resources::Responses#stream_raw} for streaming counterpart. # # Creates a model response. Provide # [text](https://platform.openai.com/docs/guides/text) or @@ -47,7 +47,7 @@ class Responses def create(params) parsed, options = OpenAI::Models::Responses::ResponseCreateParams.dump_request(params) if parsed[:stream] - message = "Please use `#create_streaming` for the streaming use case." + message = "Please use `#stream_raw` for the streaming use case." raise ArgumentError.new(message) end @client.request( @@ -73,7 +73,7 @@ def create(params) # [file search](https://platform.openai.com/docs/guides/tools-file-search) to use # your own data as input for the model's response. # - # @overload create_streaming(input:, model:, include: nil, instructions: nil, max_output_tokens: nil, metadata: nil, parallel_tool_calls: nil, previous_response_id: nil, reasoning: nil, store: nil, temperature: nil, text: nil, tool_choice: nil, tools: nil, top_p: nil, truncation: nil, user: nil, request_options: {}) + # @overload stream_raw(input:, model:, include: nil, instructions: nil, max_output_tokens: nil, metadata: nil, parallel_tool_calls: nil, previous_response_id: nil, reasoning: nil, store: nil, temperature: nil, text: nil, tool_choice: nil, tools: nil, top_p: nil, truncation: nil, user: nil, request_options: {}) # # @param input [String, Array] # @param model [String, Symbol, OpenAI::Models::ChatModel, OpenAI::Models::ResponsesModel::ResponsesOnlyModel] @@ -97,7 +97,7 @@ def create(params) # @return [OpenAI::Internal::Stream] # # @see OpenAI::Models::Responses::ResponseCreateParams - def create_streaming(params) + def stream_raw(params) parsed, options = OpenAI::Models::Responses::ResponseCreateParams.dump_request(params) unless parsed.fetch(:stream, true) message = "Please use `#create` for the non-streaming use case." diff --git a/rbi/lib/openai/resources/beta/threads.rbi b/rbi/lib/openai/resources/beta/threads.rbi index 2139613a..316bbca4 100644 --- a/rbi/lib/openai/resources/beta/threads.rbi +++ b/rbi/lib/openai/resources/beta/threads.rbi @@ -91,8 +91,7 @@ module OpenAI thread_id, request_options: {} ); end - # See {OpenAI::Resources::Beta::Threads#create_and_run_streaming} for streaming - # counterpart. + # See {OpenAI::Resources::Beta::Threads#stream_raw} for streaming counterpart. # # Create a thread and run it in one request. sig do @@ -231,8 +230,8 @@ module OpenAI # Controls for how a thread will be truncated prior to the run. Use this to # control the intial context window of the run. truncation_strategy: nil, - # There is no need to provide `stream:`. Instead, use `#create_and_run_streaming` - # or `#create_and_run` for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#stream_raw` or + # `#create_and_run` for streaming and non-streaming use cases, respectively. stream: false, request_options: {} ); end @@ -316,7 +315,7 @@ module OpenAI ] ) end - def create_and_run_streaming( + def stream_raw( # The ID of the # [assistant](https://platform.openai.com/docs/api-reference/assistants) to use to # execute this run. @@ -405,8 +404,8 @@ module OpenAI # Controls for how a thread will be truncated prior to the run. Use this to # control the intial context window of the run. truncation_strategy: nil, - # There is no need to provide `stream:`. Instead, use `#create_and_run_streaming` - # or `#create_and_run` for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#stream_raw` or + # `#create_and_run` for streaming and non-streaming use cases, respectively. stream: true, request_options: {} ); end diff --git a/rbi/lib/openai/resources/beta/threads/runs.rbi b/rbi/lib/openai/resources/beta/threads/runs.rbi index 0b26ef36..e3a1216a 100644 --- a/rbi/lib/openai/resources/beta/threads/runs.rbi +++ b/rbi/lib/openai/resources/beta/threads/runs.rbi @@ -8,7 +8,7 @@ module OpenAI sig { returns(OpenAI::Resources::Beta::Threads::Runs::Steps) } attr_reader :steps - # See {OpenAI::Resources::Beta::Threads::Runs#create_streaming} for streaming + # See {OpenAI::Resources::Beta::Threads::Runs#create_stream_raw} for streaming # counterpart. # # Create a run. @@ -173,7 +173,7 @@ module OpenAI # Body param: Controls for how a thread will be truncated prior to the run. Use # this to control the intial context window of the run. truncation_strategy: nil, - # There is no need to provide `stream:`. Instead, use `#create_streaming` or + # There is no need to provide `stream:`. Instead, use `#create_stream_raw` or # `#create` for streaming and non-streaming use cases, respectively. stream: false, request_options: {} @@ -263,7 +263,7 @@ module OpenAI ] ) end - def create_streaming( + def create_stream_raw( # Path param: The ID of the thread to run. thread_id, # Body param: The ID of the @@ -372,7 +372,7 @@ module OpenAI # Body param: Controls for how a thread will be truncated prior to the run. Use # this to control the intial context window of the run. truncation_strategy: nil, - # There is no need to provide `stream:`. Instead, use `#create_streaming` or + # There is no need to provide `stream:`. Instead, use `#create_stream_raw` or # `#create` for streaming and non-streaming use cases, respectively. stream: true, request_options: {} @@ -468,7 +468,7 @@ module OpenAI thread_id:, request_options: {} ); end - # See {OpenAI::Resources::Beta::Threads::Runs#submit_tool_outputs_streaming} for + # See {OpenAI::Resources::Beta::Threads::Runs#submit_tool_outputs_stream_raw} for # streaming counterpart. # # When a run has the `status: "requires_action"` and `required_action.type` is @@ -495,7 +495,7 @@ module OpenAI # Body param: A list of tools for which the outputs are being submitted. tool_outputs:, # There is no need to provide `stream:`. Instead, use - # `#submit_tool_outputs_streaming` or `#submit_tool_outputs` for streaming and + # `#submit_tool_outputs_stream_raw` or `#submit_tool_outputs` for streaming and # non-streaming use cases, respectively. stream: false, request_options: {} @@ -546,7 +546,7 @@ module OpenAI ] ) end - def submit_tool_outputs_streaming( + def submit_tool_outputs_stream_raw( # Path param: The ID of the run that requires the tool output submission. run_id, # Path param: The ID of the @@ -556,7 +556,7 @@ module OpenAI # Body param: A list of tools for which the outputs are being submitted. tool_outputs:, # There is no need to provide `stream:`. Instead, use - # `#submit_tool_outputs_streaming` or `#submit_tool_outputs` for streaming and + # `#submit_tool_outputs_stream_raw` or `#submit_tool_outputs` for streaming and # non-streaming use cases, respectively. stream: true, request_options: {} diff --git a/rbi/lib/openai/resources/chat/completions.rbi b/rbi/lib/openai/resources/chat/completions.rbi index d0623009..672459b2 100644 --- a/rbi/lib/openai/resources/chat/completions.rbi +++ b/rbi/lib/openai/resources/chat/completions.rbi @@ -7,8 +7,7 @@ module OpenAI sig { returns(OpenAI::Resources::Chat::Completions::Messages) } attr_reader :messages - # See {OpenAI::Resources::Chat::Completions#create_streaming} for streaming - # counterpart. + # See {OpenAI::Resources::Chat::Completions#stream_raw} for streaming counterpart. # # **Starting a new project?** We recommend trying # [Responses](https://platform.openai.com/docs/api-reference/responses) to take @@ -271,8 +270,8 @@ module OpenAI # about the # [web search tool](https://platform.openai.com/docs/guides/tools-web-search?api-mode=chat). web_search_options: nil, - # There is no need to provide `stream:`. Instead, use `#create_streaming` or - # `#create` for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#stream_raw` or `#create` + # for streaming and non-streaming use cases, respectively. stream: false, request_options: {} ); end @@ -355,7 +354,7 @@ module OpenAI ) .returns(OpenAI::Internal::Stream[OpenAI::Models::Chat::ChatCompletionChunk]) end - def create_streaming( + def stream_raw( # A list of messages comprising the conversation so far. Depending on the # [model](https://platform.openai.com/docs/models) you use, different message # types (modalities) are supported, like @@ -539,8 +538,8 @@ module OpenAI # about the # [web search tool](https://platform.openai.com/docs/guides/tools-web-search?api-mode=chat). web_search_options: nil, - # There is no need to provide `stream:`. Instead, use `#create_streaming` or - # `#create` for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#stream_raw` or `#create` + # for streaming and non-streaming use cases, respectively. stream: true, request_options: {} ); end diff --git a/rbi/lib/openai/resources/responses.rbi b/rbi/lib/openai/resources/responses.rbi index b98140db..36919e9b 100644 --- a/rbi/lib/openai/resources/responses.rbi +++ b/rbi/lib/openai/resources/responses.rbi @@ -6,7 +6,7 @@ module OpenAI sig { returns(OpenAI::Resources::Responses::InputItems) } attr_reader :input_items - # See {OpenAI::Resources::Responses#create_streaming} for streaming counterpart. + # See {OpenAI::Resources::Responses#stream_raw} for streaming counterpart. # # Creates a model response. Provide # [text](https://platform.openai.com/docs/guides/text) or @@ -183,8 +183,8 @@ module OpenAI # and detect abuse. # [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids). user: nil, - # There is no need to provide `stream:`. Instead, use `#create_streaming` or - # `#create` for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#stream_raw` or `#create` + # for streaming and non-streaming use cases, respectively. stream: false, request_options: {} ); end @@ -297,7 +297,7 @@ module OpenAI ] ) end - def create_streaming( + def stream_raw( # Text, image, or file inputs to the model, used to generate a response. # # Learn more: @@ -402,8 +402,8 @@ module OpenAI # and detect abuse. # [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids). user: nil, - # There is no need to provide `stream:`. Instead, use `#create_streaming` or - # `#create` for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#stream_raw` or `#create` + # for streaming and non-streaming use cases, respectively. stream: true, request_options: {} ); end diff --git a/sig/openai/resources/beta/threads.rbs b/sig/openai/resources/beta/threads.rbs index fe3d90cd..50be0608 100644 --- a/sig/openai/resources/beta/threads.rbs +++ b/sig/openai/resources/beta/threads.rbs @@ -49,7 +49,7 @@ module OpenAI ?request_options: OpenAI::request_opts ) -> OpenAI::Models::Beta::Threads::Run - def create_and_run_streaming: ( + def stream_raw: ( assistant_id: String, ?instructions: String?, ?max_completion_tokens: Integer?, diff --git a/sig/openai/resources/beta/threads/runs.rbs b/sig/openai/resources/beta/threads/runs.rbs index bdc66eb1..df8265f2 100644 --- a/sig/openai/resources/beta/threads/runs.rbs +++ b/sig/openai/resources/beta/threads/runs.rbs @@ -27,7 +27,7 @@ module OpenAI ?request_options: OpenAI::request_opts ) -> OpenAI::Models::Beta::Threads::Run - def create_streaming: ( + def create_stream_raw: ( String thread_id, assistant_id: String, ?include: ::Array[OpenAI::Models::Beta::Threads::Runs::run_step_include], @@ -84,7 +84,7 @@ module OpenAI ?request_options: OpenAI::request_opts ) -> OpenAI::Models::Beta::Threads::Run - def submit_tool_outputs_streaming: ( + def submit_tool_outputs_stream_raw: ( String run_id, thread_id: String, tool_outputs: ::Array[OpenAI::Models::Beta::Threads::RunSubmitToolOutputsParams::ToolOutput], diff --git a/sig/openai/resources/chat/completions.rbs b/sig/openai/resources/chat/completions.rbs index e1f27903..0e8a88ef 100644 --- a/sig/openai/resources/chat/completions.rbs +++ b/sig/openai/resources/chat/completions.rbs @@ -38,7 +38,7 @@ module OpenAI ?request_options: OpenAI::request_opts ) -> OpenAI::Models::Chat::ChatCompletion - def create_streaming: ( + def stream_raw: ( messages: ::Array[OpenAI::Models::Chat::chat_completion_message_param], model: OpenAI::Models::Chat::CompletionCreateParams::model, ?audio: OpenAI::Models::Chat::ChatCompletionAudioParam?, diff --git a/sig/openai/resources/responses.rbs b/sig/openai/resources/responses.rbs index 028c6cc6..2dfcf44d 100644 --- a/sig/openai/resources/responses.rbs +++ b/sig/openai/resources/responses.rbs @@ -24,7 +24,7 @@ module OpenAI ?request_options: OpenAI::request_opts ) -> OpenAI::Models::Responses::Response - def create_streaming: ( + def stream_raw: ( input: OpenAI::Models::Responses::ResponseCreateParams::input, model: OpenAI::Models::responses_model, ?include: ::Array[OpenAI::Models::Responses::response_includable]?,