diff --git a/.stats.yml b/.stats.yml index b21d5dae..c16913f8 100644 --- a/.stats.yml +++ b/.stats.yml @@ -1,4 +1,4 @@ configured_endpoints: 80 -openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/openai%2Fopenai-6663c59193eb95b201e492de17dcbd5e126ba03d18ce66287a3e2c632ca56fe7.yml +openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/openai%2Fopenai-4bce8217a697c729ac98046d4caf2c9e826b54c427fb0ab4f98e549a2e0ce31c.yml openapi_spec_hash: 7996d2c34cc44fe2ce9ffe93c0ab774e -config_hash: 2daae06cc598821ccf87201de0861e40 +config_hash: 31a12443afeef2933b34e2de23c40954 diff --git a/README.md b/README.md index 98c8fe21..316ac3f6 100644 --- a/README.md +++ b/README.md @@ -67,7 +67,7 @@ end We provide support for streaming responses using Server Side Events (SSE). ```ruby -stream = openai.chat.completions.create_streaming( +stream = openai.chat.completions.stream_raw( messages: [{ role: "user", content: "Say this is a test" diff --git a/lib/openai/models/audio/transcription_create_params.rb b/lib/openai/models/audio/transcription_create_params.rb index 9326a51f..ee1640c6 100644 --- a/lib/openai/models/audio/transcription_create_params.rb +++ b/lib/openai/models/audio/transcription_create_params.rb @@ -5,7 +5,7 @@ module Models module Audio # @see OpenAI::Resources::Audio::Transcriptions#create # - # @see OpenAI::Resources::Audio::Transcriptions#create_streaming + # @see OpenAI::Resources::Audio::Transcriptions#stream_raw class TranscriptionCreateParams < OpenAI::BaseModel # @!parse # extend OpenAI::Type::RequestParameters::Converter diff --git a/lib/openai/models/audio/transcription_create_response.rb b/lib/openai/models/audio/transcription_create_response.rb index be3c624c..efebb7c6 100644 --- a/lib/openai/models/audio/transcription_create_response.rb +++ b/lib/openai/models/audio/transcription_create_response.rb @@ -8,7 +8,7 @@ module Audio # # @see OpenAI::Resources::Audio::Transcriptions#create # - # @see OpenAI::Resources::Audio::Transcriptions#create_streaming + # @see OpenAI::Resources::Audio::Transcriptions#stream_raw module TranscriptionCreateResponse extend OpenAI::Union diff --git a/lib/openai/models/beta/thread_create_and_run_params.rb b/lib/openai/models/beta/thread_create_and_run_params.rb index d1a26adb..495e3cd1 100644 --- a/lib/openai/models/beta/thread_create_and_run_params.rb +++ b/lib/openai/models/beta/thread_create_and_run_params.rb @@ -5,7 +5,7 @@ module Models module Beta # @see OpenAI::Resources::Beta::Threads#create_and_run # - # @see OpenAI::Resources::Beta::Threads#create_and_run_streaming + # @see OpenAI::Resources::Beta::Threads#stream_raw class ThreadCreateAndRunParams < OpenAI::BaseModel # @!parse # extend OpenAI::Type::RequestParameters::Converter diff --git a/lib/openai/models/beta/threads/run.rb b/lib/openai/models/beta/threads/run.rb index 66e2e6a8..9b21a4a0 100644 --- a/lib/openai/models/beta/threads/run.rb +++ b/lib/openai/models/beta/threads/run.rb @@ -6,7 +6,7 @@ module Beta module Threads # @see OpenAI::Resources::Beta::Threads::Runs#create # - # @see OpenAI::Resources::Beta::Threads::Runs#create_streaming + # @see OpenAI::Resources::Beta::Threads::Runs#stream_raw class Run < OpenAI::BaseModel # @!attribute id # The identifier, which can be referenced in API endpoints. diff --git a/lib/openai/models/beta/threads/run_create_params.rb b/lib/openai/models/beta/threads/run_create_params.rb index d0c690be..a52ec57a 100644 --- a/lib/openai/models/beta/threads/run_create_params.rb +++ b/lib/openai/models/beta/threads/run_create_params.rb @@ -6,7 +6,7 @@ module Beta module Threads # @see OpenAI::Resources::Beta::Threads::Runs#create # - # @see OpenAI::Resources::Beta::Threads::Runs#create_streaming + # @see OpenAI::Resources::Beta::Threads::Runs#stream_raw class RunCreateParams < OpenAI::BaseModel # @!parse # extend OpenAI::Type::RequestParameters::Converter diff --git a/lib/openai/models/beta/threads/run_submit_tool_outputs_params.rb b/lib/openai/models/beta/threads/run_submit_tool_outputs_params.rb index 8ed57b07..1ab6c2f2 100644 --- a/lib/openai/models/beta/threads/run_submit_tool_outputs_params.rb +++ b/lib/openai/models/beta/threads/run_submit_tool_outputs_params.rb @@ -6,7 +6,7 @@ module Beta module Threads # @see OpenAI::Resources::Beta::Threads::Runs#submit_tool_outputs # - # @see OpenAI::Resources::Beta::Threads::Runs#submit_tool_outputs_streaming + # @see OpenAI::Resources::Beta::Threads::Runs#stream_raw class RunSubmitToolOutputsParams < OpenAI::BaseModel # @!parse # extend OpenAI::Type::RequestParameters::Converter diff --git a/lib/openai/models/chat/chat_completion.rb b/lib/openai/models/chat/chat_completion.rb index 2dcdfd05..5c66a291 100644 --- a/lib/openai/models/chat/chat_completion.rb +++ b/lib/openai/models/chat/chat_completion.rb @@ -5,7 +5,7 @@ module Models module Chat # @see OpenAI::Resources::Chat::Completions#create # - # @see OpenAI::Resources::Chat::Completions#create_streaming + # @see OpenAI::Resources::Chat::Completions#stream_raw class ChatCompletion < OpenAI::BaseModel # @!attribute id # A unique identifier for the chat completion. diff --git a/lib/openai/models/chat/completion_create_params.rb b/lib/openai/models/chat/completion_create_params.rb index b612cb26..f5cf3bd0 100644 --- a/lib/openai/models/chat/completion_create_params.rb +++ b/lib/openai/models/chat/completion_create_params.rb @@ -5,7 +5,7 @@ module Models module Chat # @see OpenAI::Resources::Chat::Completions#create # - # @see OpenAI::Resources::Chat::Completions#create_streaming + # @see OpenAI::Resources::Chat::Completions#stream_raw class CompletionCreateParams < OpenAI::BaseModel # @!parse # extend OpenAI::Type::RequestParameters::Converter diff --git a/lib/openai/models/completion.rb b/lib/openai/models/completion.rb index 8e3ad52a..791a65a3 100644 --- a/lib/openai/models/completion.rb +++ b/lib/openai/models/completion.rb @@ -4,7 +4,7 @@ module OpenAI module Models # @see OpenAI::Resources::Completions#create # - # @see OpenAI::Resources::Completions#create_streaming + # @see OpenAI::Resources::Completions#stream_raw class Completion < OpenAI::BaseModel # @!attribute id # A unique identifier for the completion. diff --git a/lib/openai/models/completion_create_params.rb b/lib/openai/models/completion_create_params.rb index 974ee482..cc824656 100644 --- a/lib/openai/models/completion_create_params.rb +++ b/lib/openai/models/completion_create_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models # @see OpenAI::Resources::Completions#create # - # @see OpenAI::Resources::Completions#create_streaming + # @see OpenAI::Resources::Completions#stream_raw class CompletionCreateParams < OpenAI::BaseModel # @!parse # extend OpenAI::Type::RequestParameters::Converter diff --git a/lib/openai/models/responses/response.rb b/lib/openai/models/responses/response.rb index f53f3fc5..b6e167a0 100644 --- a/lib/openai/models/responses/response.rb +++ b/lib/openai/models/responses/response.rb @@ -5,7 +5,7 @@ module Models module Responses # @see OpenAI::Resources::Responses#create # - # @see OpenAI::Resources::Responses#create_streaming + # @see OpenAI::Resources::Responses#stream_raw class Response < OpenAI::BaseModel # @!attribute id # Unique identifier for this Response. diff --git a/lib/openai/models/responses/response_create_params.rb b/lib/openai/models/responses/response_create_params.rb index 29abdd2c..55ec73cc 100644 --- a/lib/openai/models/responses/response_create_params.rb +++ b/lib/openai/models/responses/response_create_params.rb @@ -5,7 +5,7 @@ module Models module Responses # @see OpenAI::Resources::Responses#create # - # @see OpenAI::Resources::Responses#create_streaming + # @see OpenAI::Resources::Responses#stream_raw class ResponseCreateParams < OpenAI::BaseModel # @!parse # extend OpenAI::Type::RequestParameters::Converter diff --git a/lib/openai/resources/audio/transcriptions.rb b/lib/openai/resources/audio/transcriptions.rb index c9bf8f4f..2dc49631 100644 --- a/lib/openai/resources/audio/transcriptions.rb +++ b/lib/openai/resources/audio/transcriptions.rb @@ -24,7 +24,7 @@ class Transcriptions def create(params) parsed, options = OpenAI::Models::Audio::TranscriptionCreateParams.dump_request(params) if parsed[:stream] - message = "Please use `#create_streaming` for the streaming use case." + message = "Please use `#stream_raw` for the streaming use case." raise ArgumentError.new(message) end @client.request( @@ -39,7 +39,7 @@ def create(params) # Transcribes audio into the input language. # - # @overload create_streaming(file:, model:, include: nil, language: nil, prompt: nil, response_format: nil, temperature: nil, timestamp_granularities: nil, request_options: {}) + # @overload stream_raw(file:, model:, include: nil, language: nil, prompt: nil, response_format: nil, temperature: nil, timestamp_granularities: nil, request_options: {}) # # @param file [IO, StringIO] # @param model [String, Symbol, OpenAI::Models::AudioModel] @@ -54,7 +54,7 @@ def create(params) # @return [OpenAI::Stream] # # @see OpenAI::Models::Audio::TranscriptionCreateParams - def create_streaming(params) + def stream_raw(params) parsed, options = OpenAI::Models::Audio::TranscriptionCreateParams.dump_request(params) unless parsed.fetch(:stream, true) message = "Please use `#create` for the non-streaming use case." diff --git a/lib/openai/resources/beta/threads.rb b/lib/openai/resources/beta/threads.rb index 6b028b1d..82c8c329 100644 --- a/lib/openai/resources/beta/threads.rb +++ b/lib/openai/resources/beta/threads.rb @@ -121,7 +121,7 @@ def delete(thread_id, params = {}) def create_and_run(params) parsed, options = OpenAI::Models::Beta::ThreadCreateAndRunParams.dump_request(params) if parsed[:stream] - message = "Please use `#create_and_run_streaming` for the streaming use case." + message = "Please use `#stream_raw` for the streaming use case." raise ArgumentError.new(message) end @client.request( @@ -135,7 +135,7 @@ def create_and_run(params) # Create a thread and run it in one request. # - # @overload create_and_run_streaming(assistant_id:, instructions: nil, max_completion_tokens: nil, max_prompt_tokens: nil, metadata: nil, model: nil, parallel_tool_calls: nil, response_format: nil, temperature: nil, thread: nil, tool_choice: nil, tool_resources: nil, tools: nil, top_p: nil, truncation_strategy: nil, request_options: {}) + # @overload stream_raw(assistant_id:, instructions: nil, max_completion_tokens: nil, max_prompt_tokens: nil, metadata: nil, model: nil, parallel_tool_calls: nil, response_format: nil, temperature: nil, thread: nil, tool_choice: nil, tool_resources: nil, tools: nil, top_p: nil, truncation_strategy: nil, request_options: {}) # # @param assistant_id [String] # @param instructions [String, nil] @@ -157,7 +157,7 @@ def create_and_run(params) # @return [OpenAI::Stream] # # @see OpenAI::Models::Beta::ThreadCreateAndRunParams - def create_and_run_streaming(params) + def stream_raw(params) parsed, options = OpenAI::Models::Beta::ThreadCreateAndRunParams.dump_request(params) unless parsed.fetch(:stream, true) message = "Please use `#create_and_run` for the non-streaming use case." diff --git a/lib/openai/resources/beta/threads/runs.rb b/lib/openai/resources/beta/threads/runs.rb index 45bfd146..d662fbb2 100644 --- a/lib/openai/resources/beta/threads/runs.rb +++ b/lib/openai/resources/beta/threads/runs.rb @@ -38,7 +38,7 @@ class Runs def create(thread_id, params) parsed, options = OpenAI::Models::Beta::Threads::RunCreateParams.dump_request(params) if parsed[:stream] - message = "Please use `#create_streaming` for the streaming use case." + message = "Please use `#stream_raw` for the streaming use case." raise ArgumentError.new(message) end query_params = [:include] @@ -54,7 +54,7 @@ def create(thread_id, params) # Create a run. # - # @overload create_streaming(thread_id, assistant_id:, include: nil, additional_instructions: nil, additional_messages: nil, instructions: nil, max_completion_tokens: nil, max_prompt_tokens: nil, metadata: nil, model: nil, parallel_tool_calls: nil, reasoning_effort: nil, response_format: nil, temperature: nil, tool_choice: nil, tools: nil, top_p: nil, truncation_strategy: nil, request_options: {}) + # @overload stream_raw(thread_id, assistant_id:, include: nil, additional_instructions: nil, additional_messages: nil, instructions: nil, max_completion_tokens: nil, max_prompt_tokens: nil, metadata: nil, model: nil, parallel_tool_calls: nil, reasoning_effort: nil, response_format: nil, temperature: nil, tool_choice: nil, tools: nil, top_p: nil, truncation_strategy: nil, request_options: {}) # # @param thread_id [String] # @param assistant_id [String] @@ -79,7 +79,7 @@ def create(thread_id, params) # @return [OpenAI::Stream] # # @see OpenAI::Models::Beta::Threads::RunCreateParams - def create_streaming(thread_id, params) + def stream_raw(thread_id, params) parsed, options = OpenAI::Models::Beta::Threads::RunCreateParams.dump_request(params) unless parsed.fetch(:stream, true) message = "Please use `#create` for the non-streaming use case." @@ -220,7 +220,7 @@ def cancel(run_id, params) def submit_tool_outputs(run_id, params) parsed, options = OpenAI::Models::Beta::Threads::RunSubmitToolOutputsParams.dump_request(params) if parsed[:stream] - message = "Please use `#submit_tool_outputs_streaming` for the streaming use case." + message = "Please use `#stream_raw` for the streaming use case." raise ArgumentError.new(message) end thread_id = @@ -241,7 +241,7 @@ def submit_tool_outputs(run_id, params) # tool calls once they're all completed. All outputs must be submitted in a single # request. # - # @overload submit_tool_outputs_streaming(run_id, thread_id:, tool_outputs:, request_options: {}) + # @overload stream_raw(run_id, thread_id:, tool_outputs:, request_options: {}) # # @param run_id [String] # @param thread_id [String] @@ -251,7 +251,7 @@ def submit_tool_outputs(run_id, params) # @return [OpenAI::Stream] # # @see OpenAI::Models::Beta::Threads::RunSubmitToolOutputsParams - def submit_tool_outputs_streaming(run_id, params) + def stream_raw(run_id, params) parsed, options = OpenAI::Models::Beta::Threads::RunSubmitToolOutputsParams.dump_request(params) unless parsed.fetch(:stream, true) message = "Please use `#submit_tool_outputs` for the non-streaming use case." diff --git a/lib/openai/resources/chat/completions.rb b/lib/openai/resources/chat/completions.rb index 41ffd171..c000fad2 100644 --- a/lib/openai/resources/chat/completions.rb +++ b/lib/openai/resources/chat/completions.rb @@ -65,7 +65,7 @@ class Completions def create(params) parsed, options = OpenAI::Models::Chat::CompletionCreateParams.dump_request(params) if parsed[:stream] - message = "Please use `#create_streaming` for the streaming use case." + message = "Please use `#stream_raw` for the streaming use case." raise ArgumentError.new(message) end @client.request( @@ -95,7 +95,7 @@ def create(params) # unsupported parameters in reasoning models, # [refer to the reasoning guide](https://platform.openai.com/docs/guides/reasoning). # - # @overload create_streaming(messages:, model:, audio: nil, frequency_penalty: nil, function_call: nil, functions: nil, logit_bias: nil, logprobs: nil, max_completion_tokens: nil, max_tokens: nil, metadata: nil, modalities: nil, n: nil, parallel_tool_calls: nil, prediction: nil, presence_penalty: nil, reasoning_effort: nil, response_format: nil, seed: nil, service_tier: nil, stop: nil, store: nil, stream_options: nil, temperature: nil, tool_choice: nil, tools: nil, top_logprobs: nil, top_p: nil, user: nil, web_search_options: nil, request_options: {}) + # @overload stream_raw(messages:, model:, audio: nil, frequency_penalty: nil, function_call: nil, functions: nil, logit_bias: nil, logprobs: nil, max_completion_tokens: nil, max_tokens: nil, metadata: nil, modalities: nil, n: nil, parallel_tool_calls: nil, prediction: nil, presence_penalty: nil, reasoning_effort: nil, response_format: nil, seed: nil, service_tier: nil, stop: nil, store: nil, stream_options: nil, temperature: nil, tool_choice: nil, tools: nil, top_logprobs: nil, top_p: nil, user: nil, web_search_options: nil, request_options: {}) # # @param messages [Array] # @param model [String, Symbol, OpenAI::Models::ChatModel] @@ -132,7 +132,7 @@ def create(params) # @return [OpenAI::Stream] # # @see OpenAI::Models::Chat::CompletionCreateParams - def create_streaming(params) + def stream_raw(params) parsed, options = OpenAI::Models::Chat::CompletionCreateParams.dump_request(params) unless parsed.fetch(:stream, true) message = "Please use `#create` for the non-streaming use case." diff --git a/lib/openai/resources/completions.rb b/lib/openai/resources/completions.rb index f27ec1e4..62a26a53 100644 --- a/lib/openai/resources/completions.rb +++ b/lib/openai/resources/completions.rb @@ -32,7 +32,7 @@ class Completions def create(params) parsed, options = OpenAI::Models::CompletionCreateParams.dump_request(params) if parsed[:stream] - message = "Please use `#create_streaming` for the streaming use case." + message = "Please use `#stream_raw` for the streaming use case." raise ArgumentError.new(message) end @client.request( @@ -46,7 +46,7 @@ def create(params) # Creates a completion for the provided prompt and parameters. # - # @overload create_streaming(model:, prompt:, best_of: nil, echo: nil, frequency_penalty: nil, logit_bias: nil, logprobs: nil, max_tokens: nil, n: nil, presence_penalty: nil, seed: nil, stop: nil, stream_options: nil, suffix: nil, temperature: nil, top_p: nil, user: nil, request_options: {}) + # @overload stream_raw(model:, prompt:, best_of: nil, echo: nil, frequency_penalty: nil, logit_bias: nil, logprobs: nil, max_tokens: nil, n: nil, presence_penalty: nil, seed: nil, stop: nil, stream_options: nil, suffix: nil, temperature: nil, top_p: nil, user: nil, request_options: {}) # # @param model [String, Symbol, OpenAI::Models::CompletionCreateParams::Model] # @param prompt [String, Array, Array, Array>, nil] @@ -70,7 +70,7 @@ def create(params) # @return [OpenAI::Stream] # # @see OpenAI::Models::CompletionCreateParams - def create_streaming(params) + def stream_raw(params) parsed, options = OpenAI::Models::CompletionCreateParams.dump_request(params) unless parsed.fetch(:stream, true) message = "Please use `#create` for the non-streaming use case." diff --git a/lib/openai/resources/responses.rb b/lib/openai/resources/responses.rb index 2e39f489..46de5e29 100644 --- a/lib/openai/resources/responses.rb +++ b/lib/openai/resources/responses.rb @@ -45,7 +45,7 @@ class Responses def create(params) parsed, options = OpenAI::Models::Responses::ResponseCreateParams.dump_request(params) if parsed[:stream] - message = "Please use `#create_streaming` for the streaming use case." + message = "Please use `#stream_raw` for the streaming use case." raise ArgumentError.new(message) end @client.request( @@ -69,7 +69,7 @@ def create(params) # [file search](https://platform.openai.com/docs/guides/tools-file-search) to use # your own data as input for the model's response. # - # @overload create_streaming(input:, model:, include: nil, instructions: nil, max_output_tokens: nil, metadata: nil, parallel_tool_calls: nil, previous_response_id: nil, reasoning: nil, store: nil, temperature: nil, text: nil, tool_choice: nil, tools: nil, top_p: nil, truncation: nil, user: nil, request_options: {}) + # @overload stream_raw(input:, model:, include: nil, instructions: nil, max_output_tokens: nil, metadata: nil, parallel_tool_calls: nil, previous_response_id: nil, reasoning: nil, store: nil, temperature: nil, text: nil, tool_choice: nil, tools: nil, top_p: nil, truncation: nil, user: nil, request_options: {}) # # @param input [String, Array] # @param model [String, Symbol, OpenAI::Models::ChatModel, OpenAI::Models::ResponsesModel] @@ -93,7 +93,7 @@ def create(params) # @return [OpenAI::Stream] # # @see OpenAI::Models::Responses::ResponseCreateParams - def create_streaming(params) + def stream_raw(params) parsed, options = OpenAI::Models::Responses::ResponseCreateParams.dump_request(params) unless parsed.fetch(:stream, true) message = "Please use `#create` for the non-streaming use case." diff --git a/rbi/lib/openai/resources/audio/transcriptions.rbi b/rbi/lib/openai/resources/audio/transcriptions.rbi index 688f6c9f..476b325a 100644 --- a/rbi/lib/openai/resources/audio/transcriptions.rbi +++ b/rbi/lib/openai/resources/audio/transcriptions.rbi @@ -59,8 +59,8 @@ module OpenAI # is no additional latency for segment timestamps, but generating word timestamps # incurs additional latency. timestamp_granularities: nil, - # There is no need to provide `stream:`. Instead, use `#create_streaming` or - # `#create` for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#stream_raw` or `#create` + # for streaming and non-streaming use cases, respectively. stream: false, request_options: {} ) @@ -89,7 +89,7 @@ module OpenAI ] ) end - def create_streaming( + def stream_raw( # The audio file object (not file name) to transcribe, in one of these formats: # flac, mp3, mp4, mpeg, mpga, m4a, ogg, wav, or webm. file:, @@ -128,8 +128,8 @@ module OpenAI # is no additional latency for segment timestamps, but generating word timestamps # incurs additional latency. timestamp_granularities: nil, - # There is no need to provide `stream:`. Instead, use `#create_streaming` or - # `#create` for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#stream_raw` or `#create` + # for streaming and non-streaming use cases, respectively. stream: true, request_options: {} ) diff --git a/rbi/lib/openai/resources/beta/threads.rbi b/rbi/lib/openai/resources/beta/threads.rbi index d5a0da4b..114b3406 100644 --- a/rbi/lib/openai/resources/beta/threads.rbi +++ b/rbi/lib/openai/resources/beta/threads.rbi @@ -236,8 +236,8 @@ module OpenAI # Controls for how a thread will be truncated prior to the run. Use this to # control the intial context window of the run. truncation_strategy: nil, - # There is no need to provide `stream:`. Instead, use `#create_and_run_streaming` - # or `#create_and_run` for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#stream_raw` or + # `#create_and_run` for streaming and non-streaming use cases, respectively. stream: false, request_options: {} ) @@ -320,7 +320,7 @@ module OpenAI ] ) end - def create_and_run_streaming( + def stream_raw( # The ID of the # [assistant](https://platform.openai.com/docs/api-reference/assistants) to use to # execute this run. @@ -409,8 +409,8 @@ module OpenAI # Controls for how a thread will be truncated prior to the run. Use this to # control the intial context window of the run. truncation_strategy: nil, - # There is no need to provide `stream:`. Instead, use `#create_and_run_streaming` - # or `#create_and_run` for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#stream_raw` or + # `#create_and_run` for streaming and non-streaming use cases, respectively. stream: true, request_options: {} ) diff --git a/rbi/lib/openai/resources/beta/threads/runs.rbi b/rbi/lib/openai/resources/beta/threads/runs.rbi index 296929f4..704f83ce 100644 --- a/rbi/lib/openai/resources/beta/threads/runs.rbi +++ b/rbi/lib/openai/resources/beta/threads/runs.rbi @@ -170,8 +170,8 @@ module OpenAI # Body param: Controls for how a thread will be truncated prior to the run. Use # this to control the intial context window of the run. truncation_strategy: nil, - # There is no need to provide `stream:`. Instead, use `#create_streaming` or - # `#create` for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#stream_raw` or `#create` + # for streaming and non-streaming use cases, respectively. stream: false, request_options: {} ) @@ -259,7 +259,7 @@ module OpenAI ] ) end - def create_streaming( + def stream_raw( # Path param: The ID of the thread to run. thread_id, # Body param: The ID of the @@ -368,8 +368,8 @@ module OpenAI # Body param: Controls for how a thread will be truncated prior to the run. Use # this to control the intial context window of the run. truncation_strategy: nil, - # There is no need to provide `stream:`. Instead, use `#create_streaming` or - # `#create` for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#stream_raw` or `#create` + # for streaming and non-streaming use cases, respectively. stream: true, request_options: {} ) @@ -497,9 +497,8 @@ module OpenAI thread_id:, # Body param: A list of tools for which the outputs are being submitted. tool_outputs:, - # There is no need to provide `stream:`. Instead, use - # `#submit_tool_outputs_streaming` or `#submit_tool_outputs` for streaming and - # non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#stream_raw` or + # `#submit_tool_outputs` for streaming and non-streaming use cases, respectively. stream: false, request_options: {} ) @@ -548,7 +547,7 @@ module OpenAI ] ) end - def submit_tool_outputs_streaming( + def stream_raw( # Path param: The ID of the run that requires the tool output submission. run_id, # Path param: The ID of the @@ -557,9 +556,8 @@ module OpenAI thread_id:, # Body param: A list of tools for which the outputs are being submitted. tool_outputs:, - # There is no need to provide `stream:`. Instead, use - # `#submit_tool_outputs_streaming` or `#submit_tool_outputs` for streaming and - # non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#stream_raw` or + # `#submit_tool_outputs` for streaming and non-streaming use cases, respectively. stream: true, request_options: {} ) diff --git a/rbi/lib/openai/resources/chat/completions.rbi b/rbi/lib/openai/resources/chat/completions.rbi index 595d5865..0471b372 100644 --- a/rbi/lib/openai/resources/chat/completions.rbi +++ b/rbi/lib/openai/resources/chat/completions.rbi @@ -268,8 +268,8 @@ module OpenAI # about the # [web search tool](https://platform.openai.com/docs/guides/tools-web-search?api-mode=chat). web_search_options: nil, - # There is no need to provide `stream:`. Instead, use `#create_streaming` or - # `#create` for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#stream_raw` or `#create` + # for streaming and non-streaming use cases, respectively. stream: false, request_options: {} ) @@ -352,7 +352,7 @@ module OpenAI ) .returns(OpenAI::Stream[OpenAI::Models::Chat::ChatCompletionChunk]) end - def create_streaming( + def stream_raw( # A list of messages comprising the conversation so far. Depending on the # [model](https://platform.openai.com/docs/models) you use, different message # types (modalities) are supported, like @@ -536,8 +536,8 @@ module OpenAI # about the # [web search tool](https://platform.openai.com/docs/guides/tools-web-search?api-mode=chat). web_search_options: nil, - # There is no need to provide `stream:`. Instead, use `#create_streaming` or - # `#create` for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#stream_raw` or `#create` + # for streaming and non-streaming use cases, respectively. stream: true, request_options: {} ) diff --git a/rbi/lib/openai/resources/completions.rbi b/rbi/lib/openai/resources/completions.rbi index b5b3feeb..7c97579b 100644 --- a/rbi/lib/openai/resources/completions.rbi +++ b/rbi/lib/openai/resources/completions.rbi @@ -139,8 +139,8 @@ module OpenAI # and detect abuse. # [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids). user: nil, - # There is no need to provide `stream:`. Instead, use `#create_streaming` or - # `#create` for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#stream_raw` or `#create` + # for streaming and non-streaming use cases, respectively. stream: false, request_options: {} ) @@ -178,7 +178,7 @@ module OpenAI ) .returns(OpenAI::Stream[OpenAI::Models::Completion]) end - def create_streaming( + def stream_raw( # ID of the model to use. You can use the # [List models](https://platform.openai.com/docs/api-reference/models/list) API to # see all of your available models, or see our @@ -282,8 +282,8 @@ module OpenAI # and detect abuse. # [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids). user: nil, - # There is no need to provide `stream:`. Instead, use `#create_streaming` or - # `#create` for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#stream_raw` or `#create` + # for streaming and non-streaming use cases, respectively. stream: true, request_options: {} ) diff --git a/rbi/lib/openai/resources/responses.rbi b/rbi/lib/openai/resources/responses.rbi index 8a9352b4..4d305614 100644 --- a/rbi/lib/openai/resources/responses.rbi +++ b/rbi/lib/openai/resources/responses.rbi @@ -177,8 +177,8 @@ module OpenAI # and detect abuse. # [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids). user: nil, - # There is no need to provide `stream:`. Instead, use `#create_streaming` or - # `#create` for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#stream_raw` or `#create` + # for streaming and non-streaming use cases, respectively. stream: false, request_options: {} ) @@ -287,7 +287,7 @@ module OpenAI ] ) end - def create_streaming( + def stream_raw( # Text, image, or file inputs to the model, used to generate a response. # # Learn more: @@ -392,8 +392,8 @@ module OpenAI # and detect abuse. # [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids). user: nil, - # There is no need to provide `stream:`. Instead, use `#create_streaming` or - # `#create` for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#stream_raw` or `#create` + # for streaming and non-streaming use cases, respectively. stream: true, request_options: {} ) diff --git a/sig/openai/resources/audio/transcriptions.rbs b/sig/openai/resources/audio/transcriptions.rbs index b52531d0..6f8acd12 100644 --- a/sig/openai/resources/audio/transcriptions.rbs +++ b/sig/openai/resources/audio/transcriptions.rbs @@ -14,7 +14,7 @@ module OpenAI ?request_options: OpenAI::request_opts ) -> OpenAI::Models::Audio::transcription_create_response - def create_streaming: ( + def stream_raw: ( file: IO | StringIO, model: OpenAI::Models::Audio::TranscriptionCreateParams::model, ?include: ::Array[OpenAI::Models::Audio::transcription_include], diff --git a/sig/openai/resources/beta/threads.rbs b/sig/openai/resources/beta/threads.rbs index 10e58839..7ee78295 100644 --- a/sig/openai/resources/beta/threads.rbs +++ b/sig/openai/resources/beta/threads.rbs @@ -49,7 +49,7 @@ module OpenAI ?request_options: OpenAI::request_opts ) -> OpenAI::Models::Beta::Threads::Run - def create_and_run_streaming: ( + def stream_raw: ( assistant_id: String, ?instructions: String?, ?max_completion_tokens: Integer?, diff --git a/sig/openai/resources/beta/threads/runs.rbs b/sig/openai/resources/beta/threads/runs.rbs index 24c1f8c5..dae2f06c 100644 --- a/sig/openai/resources/beta/threads/runs.rbs +++ b/sig/openai/resources/beta/threads/runs.rbs @@ -27,7 +27,7 @@ module OpenAI ?request_options: OpenAI::request_opts ) -> OpenAI::Models::Beta::Threads::Run - def create_streaming: ( + def stream_raw: ( String thread_id, assistant_id: String, ?include: ::Array[OpenAI::Models::Beta::Threads::Runs::run_step_include], @@ -84,7 +84,7 @@ module OpenAI ?request_options: OpenAI::request_opts ) -> OpenAI::Models::Beta::Threads::Run - def submit_tool_outputs_streaming: ( + def stream_raw: ( String run_id, thread_id: String, tool_outputs: ::Array[OpenAI::Models::Beta::Threads::RunSubmitToolOutputsParams::ToolOutput], diff --git a/sig/openai/resources/chat/completions.rbs b/sig/openai/resources/chat/completions.rbs index 5bd7a8db..b699bc30 100644 --- a/sig/openai/resources/chat/completions.rbs +++ b/sig/openai/resources/chat/completions.rbs @@ -38,7 +38,7 @@ module OpenAI ?request_options: OpenAI::request_opts ) -> OpenAI::Models::Chat::ChatCompletion - def create_streaming: ( + def stream_raw: ( messages: ::Array[OpenAI::Models::Chat::chat_completion_message_param], model: OpenAI::Models::Chat::CompletionCreateParams::model, ?audio: OpenAI::Models::Chat::ChatCompletionAudioParam?, diff --git a/sig/openai/resources/completions.rbs b/sig/openai/resources/completions.rbs index 42f91241..b48f77df 100644 --- a/sig/openai/resources/completions.rbs +++ b/sig/openai/resources/completions.rbs @@ -22,7 +22,7 @@ module OpenAI ?request_options: OpenAI::request_opts ) -> OpenAI::Models::Completion - def create_streaming: ( + def stream_raw: ( model: OpenAI::Models::CompletionCreateParams::model, prompt: OpenAI::Models::CompletionCreateParams::prompt?, ?best_of: Integer?, diff --git a/sig/openai/resources/responses.rbs b/sig/openai/resources/responses.rbs index ba80da78..51041a3b 100644 --- a/sig/openai/resources/responses.rbs +++ b/sig/openai/resources/responses.rbs @@ -24,7 +24,7 @@ module OpenAI ?request_options: OpenAI::request_opts ) -> OpenAI::Models::Responses::Response - def create_streaming: ( + def stream_raw: ( input: OpenAI::Models::Responses::ResponseCreateParams::input, model: OpenAI::Models::responses_model, ?include: ::Array[OpenAI::Models::Responses::response_includable]?,