From cd9f6d3c3608b178c5beaab7420c2caebce1a4bf Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Thu, 3 Apr 2025 18:38:39 +0000 Subject: [PATCH] feat(api): manual updates --- .stats.yml | 2 +- .../audio/transcription_create_params.rb | 2 +- .../audio/transcription_create_response.rb | 2 +- lib/openai/models/beta/threads/run.rb | 2 +- .../models/beta/threads/run_create_params.rb | 2 +- .../threads/run_submit_tool_outputs_params.rb | 2 +- lib/openai/models/completion.rb | 2 +- lib/openai/models/completion_create_params.rb | 2 +- lib/openai/resources/audio/transcriptions.rb | 6 ++--- lib/openai/resources/beta/threads/runs.rb | 12 +++++----- lib/openai/resources/completions.rb | 6 ++--- .../openai/resources/audio/transcriptions.rbi | 10 ++++----- .../openai/resources/beta/threads/runs.rbi | 22 ++++++++++--------- rbi/lib/openai/resources/completions.rbi | 10 ++++----- sig/openai/resources/audio/transcriptions.rbs | 2 +- sig/openai/resources/beta/threads/runs.rbs | 4 ++-- sig/openai/resources/completions.rbs | 2 +- 17 files changed, 46 insertions(+), 44 deletions(-) diff --git a/.stats.yml b/.stats.yml index b80d385d..e0e1a71e 100644 --- a/.stats.yml +++ b/.stats.yml @@ -1,4 +1,4 @@ configured_endpoints: 80 openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/openai%2Fopenai-4bce8217a697c729ac98046d4caf2c9e826b54c427fb0ab4f98e549a2e0ce31c.yml openapi_spec_hash: 7996d2c34cc44fe2ce9ffe93c0ab774e -config_hash: 578c5bff4208d560c0c280f13324409f +config_hash: bcd2cacdcb9fae9938f273cd167f613c diff --git a/lib/openai/models/audio/transcription_create_params.rb b/lib/openai/models/audio/transcription_create_params.rb index f3dc4d0d..0cd4367b 100644 --- a/lib/openai/models/audio/transcription_create_params.rb +++ b/lib/openai/models/audio/transcription_create_params.rb @@ -5,7 +5,7 @@ module Models module Audio # @see OpenAI::Resources::Audio::Transcriptions#create # - # @see OpenAI::Resources::Audio::Transcriptions#stream_raw + # @see OpenAI::Resources::Audio::Transcriptions#create_streaming class TranscriptionCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter diff --git a/lib/openai/models/audio/transcription_create_response.rb b/lib/openai/models/audio/transcription_create_response.rb index a8026630..75cfc88b 100644 --- a/lib/openai/models/audio/transcription_create_response.rb +++ b/lib/openai/models/audio/transcription_create_response.rb @@ -8,7 +8,7 @@ module Audio # # @see OpenAI::Resources::Audio::Transcriptions#create # - # @see OpenAI::Resources::Audio::Transcriptions#stream_raw + # @see OpenAI::Resources::Audio::Transcriptions#create_streaming module TranscriptionCreateResponse extend OpenAI::Internal::Type::Union diff --git a/lib/openai/models/beta/threads/run.rb b/lib/openai/models/beta/threads/run.rb index 8af69bee..9377b199 100644 --- a/lib/openai/models/beta/threads/run.rb +++ b/lib/openai/models/beta/threads/run.rb @@ -6,7 +6,7 @@ module Beta module Threads # @see OpenAI::Resources::Beta::Threads::Runs#create # - # @see OpenAI::Resources::Beta::Threads::Runs#stream_raw + # @see OpenAI::Resources::Beta::Threads::Runs#create_stream_raw class Run < OpenAI::Internal::Type::BaseModel # @!attribute id # The identifier, which can be referenced in API endpoints. diff --git a/lib/openai/models/beta/threads/run_create_params.rb b/lib/openai/models/beta/threads/run_create_params.rb index 51a64adc..612590b0 100644 --- a/lib/openai/models/beta/threads/run_create_params.rb +++ b/lib/openai/models/beta/threads/run_create_params.rb @@ -6,7 +6,7 @@ module Beta module Threads # @see OpenAI::Resources::Beta::Threads::Runs#create # - # @see OpenAI::Resources::Beta::Threads::Runs#stream_raw + # @see OpenAI::Resources::Beta::Threads::Runs#create_stream_raw class RunCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter diff --git a/lib/openai/models/beta/threads/run_submit_tool_outputs_params.rb b/lib/openai/models/beta/threads/run_submit_tool_outputs_params.rb index 85d3e4d9..4e4733e6 100644 --- a/lib/openai/models/beta/threads/run_submit_tool_outputs_params.rb +++ b/lib/openai/models/beta/threads/run_submit_tool_outputs_params.rb @@ -6,7 +6,7 @@ module Beta module Threads # @see OpenAI::Resources::Beta::Threads::Runs#submit_tool_outputs # - # @see OpenAI::Resources::Beta::Threads::Runs#submit_tool_stream_raw + # @see OpenAI::Resources::Beta::Threads::Runs#submit_tool_outputs_stream_raw class RunSubmitToolOutputsParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter diff --git a/lib/openai/models/completion.rb b/lib/openai/models/completion.rb index 5d7fd8b0..49f1d4c1 100644 --- a/lib/openai/models/completion.rb +++ b/lib/openai/models/completion.rb @@ -4,7 +4,7 @@ module OpenAI module Models # @see OpenAI::Resources::Completions#create # - # @see OpenAI::Resources::Completions#stream_raw + # @see OpenAI::Resources::Completions#create_streaming class Completion < OpenAI::Internal::Type::BaseModel # @!attribute id # A unique identifier for the completion. diff --git a/lib/openai/models/completion_create_params.rb b/lib/openai/models/completion_create_params.rb index e17ce7d4..63d127ca 100644 --- a/lib/openai/models/completion_create_params.rb +++ b/lib/openai/models/completion_create_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models # @see OpenAI::Resources::Completions#create # - # @see OpenAI::Resources::Completions#stream_raw + # @see OpenAI::Resources::Completions#create_streaming class CompletionCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter diff --git a/lib/openai/resources/audio/transcriptions.rb b/lib/openai/resources/audio/transcriptions.rb index 677b18f1..bd918fe4 100644 --- a/lib/openai/resources/audio/transcriptions.rb +++ b/lib/openai/resources/audio/transcriptions.rb @@ -24,7 +24,7 @@ class Transcriptions def create(params) parsed, options = OpenAI::Models::Audio::TranscriptionCreateParams.dump_request(params) if parsed[:stream] - message = "Please use `#stream_raw` for the streaming use case." + message = "Please use `#create_streaming` for the streaming use case." raise ArgumentError.new(message) end @client.request( @@ -39,7 +39,7 @@ def create(params) # Transcribes audio into the input language. # - # @overload stream_raw(file:, model:, include: nil, language: nil, prompt: nil, response_format: nil, temperature: nil, timestamp_granularities: nil, request_options: {}) + # @overload create_streaming(file:, model:, include: nil, language: nil, prompt: nil, response_format: nil, temperature: nil, timestamp_granularities: nil, request_options: {}) # # @param file [IO, StringIO] # @param model [String, Symbol, OpenAI::Models::AudioModel] @@ -54,7 +54,7 @@ def create(params) # @return [OpenAI::Internal::Stream] # # @see OpenAI::Models::Audio::TranscriptionCreateParams - def stream_raw(params) + def create_streaming(params) parsed, options = OpenAI::Models::Audio::TranscriptionCreateParams.dump_request(params) unless parsed.fetch(:stream, true) message = "Please use `#create` for the non-streaming use case." diff --git a/lib/openai/resources/beta/threads/runs.rb b/lib/openai/resources/beta/threads/runs.rb index aa61011c..bce39bba 100644 --- a/lib/openai/resources/beta/threads/runs.rb +++ b/lib/openai/resources/beta/threads/runs.rb @@ -38,7 +38,7 @@ class Runs def create(thread_id, params) parsed, options = OpenAI::Models::Beta::Threads::RunCreateParams.dump_request(params) if parsed[:stream] - message = "Please use `#stream_raw` for the streaming use case." + message = "Please use `#create_stream_raw` for the streaming use case." raise ArgumentError.new(message) end query_params = [:include] @@ -54,7 +54,7 @@ def create(thread_id, params) # Create a run. # - # @overload stream_raw(thread_id, assistant_id:, include: nil, additional_instructions: nil, additional_messages: nil, instructions: nil, max_completion_tokens: nil, max_prompt_tokens: nil, metadata: nil, model: nil, parallel_tool_calls: nil, reasoning_effort: nil, response_format: nil, temperature: nil, tool_choice: nil, tools: nil, top_p: nil, truncation_strategy: nil, request_options: {}) + # @overload create_stream_raw(thread_id, assistant_id:, include: nil, additional_instructions: nil, additional_messages: nil, instructions: nil, max_completion_tokens: nil, max_prompt_tokens: nil, metadata: nil, model: nil, parallel_tool_calls: nil, reasoning_effort: nil, response_format: nil, temperature: nil, tool_choice: nil, tools: nil, top_p: nil, truncation_strategy: nil, request_options: {}) # # @param thread_id [String] # @param assistant_id [String] @@ -79,7 +79,7 @@ def create(thread_id, params) # @return [OpenAI::Internal::Stream] # # @see OpenAI::Models::Beta::Threads::RunCreateParams - def stream_raw(thread_id, params) + def create_stream_raw(thread_id, params) parsed, options = OpenAI::Models::Beta::Threads::RunCreateParams.dump_request(params) unless parsed.fetch(:stream, true) message = "Please use `#create` for the non-streaming use case." @@ -220,7 +220,7 @@ def cancel(run_id, params) def submit_tool_outputs(run_id, params) parsed, options = OpenAI::Models::Beta::Threads::RunSubmitToolOutputsParams.dump_request(params) if parsed[:stream] - message = "Please use `#submit_tool_stream_raw` for the streaming use case." + message = "Please use `#submit_tool_outputs_stream_raw` for the streaming use case." raise ArgumentError.new(message) end thread_id = @@ -241,7 +241,7 @@ def submit_tool_outputs(run_id, params) # tool calls once they're all completed. All outputs must be submitted in a single # request. # - # @overload submit_tool_stream_raw(run_id, thread_id:, tool_outputs:, request_options: {}) + # @overload submit_tool_outputs_stream_raw(run_id, thread_id:, tool_outputs:, request_options: {}) # # @param run_id [String] # @param thread_id [String] @@ -251,7 +251,7 @@ def submit_tool_outputs(run_id, params) # @return [OpenAI::Internal::Stream] # # @see OpenAI::Models::Beta::Threads::RunSubmitToolOutputsParams - def submit_tool_stream_raw(run_id, params) + def submit_tool_outputs_stream_raw(run_id, params) parsed, options = OpenAI::Models::Beta::Threads::RunSubmitToolOutputsParams.dump_request(params) unless parsed.fetch(:stream, true) message = "Please use `#submit_tool_outputs` for the non-streaming use case." diff --git a/lib/openai/resources/completions.rb b/lib/openai/resources/completions.rb index 861ab13a..06c5a7dd 100644 --- a/lib/openai/resources/completions.rb +++ b/lib/openai/resources/completions.rb @@ -32,7 +32,7 @@ class Completions def create(params) parsed, options = OpenAI::Models::CompletionCreateParams.dump_request(params) if parsed[:stream] - message = "Please use `#stream_raw` for the streaming use case." + message = "Please use `#create_streaming` for the streaming use case." raise ArgumentError.new(message) end @client.request( @@ -46,7 +46,7 @@ def create(params) # Creates a completion for the provided prompt and parameters. # - # @overload stream_raw(model:, prompt:, best_of: nil, echo: nil, frequency_penalty: nil, logit_bias: nil, logprobs: nil, max_tokens: nil, n: nil, presence_penalty: nil, seed: nil, stop: nil, stream_options: nil, suffix: nil, temperature: nil, top_p: nil, user: nil, request_options: {}) + # @overload create_streaming(model:, prompt:, best_of: nil, echo: nil, frequency_penalty: nil, logit_bias: nil, logprobs: nil, max_tokens: nil, n: nil, presence_penalty: nil, seed: nil, stop: nil, stream_options: nil, suffix: nil, temperature: nil, top_p: nil, user: nil, request_options: {}) # # @param model [String, Symbol, OpenAI::Models::CompletionCreateParams::Model] # @param prompt [String, Array, Array, Array>, nil] @@ -70,7 +70,7 @@ def create(params) # @return [OpenAI::Internal::Stream] # # @see OpenAI::Models::CompletionCreateParams - def stream_raw(params) + def create_streaming(params) parsed, options = OpenAI::Models::CompletionCreateParams.dump_request(params) unless parsed.fetch(:stream, true) message = "Please use `#create` for the non-streaming use case." diff --git a/rbi/lib/openai/resources/audio/transcriptions.rbi b/rbi/lib/openai/resources/audio/transcriptions.rbi index d0f71924..047d7619 100644 --- a/rbi/lib/openai/resources/audio/transcriptions.rbi +++ b/rbi/lib/openai/resources/audio/transcriptions.rbi @@ -59,8 +59,8 @@ module OpenAI # is no additional latency for segment timestamps, but generating word timestamps # incurs additional latency. timestamp_granularities: nil, - # There is no need to provide `stream:`. Instead, use `#stream_raw` or `#create` - # for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#create_streaming` or + # `#create` for streaming and non-streaming use cases, respectively. stream: false, request_options: {} ) @@ -89,7 +89,7 @@ module OpenAI ] ) end - def stream_raw( + def create_streaming( # The audio file object (not file name) to transcribe, in one of these formats: # flac, mp3, mp4, mpeg, mpga, m4a, ogg, wav, or webm. file:, @@ -128,8 +128,8 @@ module OpenAI # is no additional latency for segment timestamps, but generating word timestamps # incurs additional latency. timestamp_granularities: nil, - # There is no need to provide `stream:`. Instead, use `#stream_raw` or `#create` - # for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#create_streaming` or + # `#create` for streaming and non-streaming use cases, respectively. stream: true, request_options: {} ) diff --git a/rbi/lib/openai/resources/beta/threads/runs.rbi b/rbi/lib/openai/resources/beta/threads/runs.rbi index a1f66f91..68622174 100644 --- a/rbi/lib/openai/resources/beta/threads/runs.rbi +++ b/rbi/lib/openai/resources/beta/threads/runs.rbi @@ -170,8 +170,8 @@ module OpenAI # Body param: Controls for how a thread will be truncated prior to the run. Use # this to control the intial context window of the run. truncation_strategy: nil, - # There is no need to provide `stream:`. Instead, use `#stream_raw` or `#create` - # for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#create_stream_raw` or + # `#create` for streaming and non-streaming use cases, respectively. stream: false, request_options: {} ) @@ -259,7 +259,7 @@ module OpenAI ] ) end - def stream_raw( + def create_stream_raw( # Path param: The ID of the thread to run. thread_id, # Body param: The ID of the @@ -368,8 +368,8 @@ module OpenAI # Body param: Controls for how a thread will be truncated prior to the run. Use # this to control the intial context window of the run. truncation_strategy: nil, - # There is no need to provide `stream:`. Instead, use `#stream_raw` or `#create` - # for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#create_stream_raw` or + # `#create` for streaming and non-streaming use cases, respectively. stream: true, request_options: {} ) @@ -497,8 +497,9 @@ module OpenAI thread_id:, # Body param: A list of tools for which the outputs are being submitted. tool_outputs:, - # There is no need to provide `stream:`. Instead, use `#submit_tool_stream_raw` or - # `#submit_tool_outputs` for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use + # `#submit_tool_outputs_stream_raw` or `#submit_tool_outputs` for streaming and + # non-streaming use cases, respectively. stream: false, request_options: {} ) @@ -547,7 +548,7 @@ module OpenAI ] ) end - def submit_tool_stream_raw( + def submit_tool_outputs_stream_raw( # Path param: The ID of the run that requires the tool output submission. run_id, # Path param: The ID of the @@ -556,8 +557,9 @@ module OpenAI thread_id:, # Body param: A list of tools for which the outputs are being submitted. tool_outputs:, - # There is no need to provide `stream:`. Instead, use `#submit_tool_stream_raw` or - # `#submit_tool_outputs` for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use + # `#submit_tool_outputs_stream_raw` or `#submit_tool_outputs` for streaming and + # non-streaming use cases, respectively. stream: true, request_options: {} ) diff --git a/rbi/lib/openai/resources/completions.rbi b/rbi/lib/openai/resources/completions.rbi index 08f12a97..d064199d 100644 --- a/rbi/lib/openai/resources/completions.rbi +++ b/rbi/lib/openai/resources/completions.rbi @@ -139,8 +139,8 @@ module OpenAI # and detect abuse. # [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids). user: nil, - # There is no need to provide `stream:`. Instead, use `#stream_raw` or `#create` - # for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#create_streaming` or + # `#create` for streaming and non-streaming use cases, respectively. stream: false, request_options: {} ) @@ -178,7 +178,7 @@ module OpenAI ) .returns(OpenAI::Internal::Stream[OpenAI::Models::Completion]) end - def stream_raw( + def create_streaming( # ID of the model to use. You can use the # [List models](https://platform.openai.com/docs/api-reference/models/list) API to # see all of your available models, or see our @@ -282,8 +282,8 @@ module OpenAI # and detect abuse. # [Learn more](https://platform.openai.com/docs/guides/safety-best-practices#end-user-ids). user: nil, - # There is no need to provide `stream:`. Instead, use `#stream_raw` or `#create` - # for streaming and non-streaming use cases, respectively. + # There is no need to provide `stream:`. Instead, use `#create_streaming` or + # `#create` for streaming and non-streaming use cases, respectively. stream: true, request_options: {} ) diff --git a/sig/openai/resources/audio/transcriptions.rbs b/sig/openai/resources/audio/transcriptions.rbs index e1577759..6506a498 100644 --- a/sig/openai/resources/audio/transcriptions.rbs +++ b/sig/openai/resources/audio/transcriptions.rbs @@ -14,7 +14,7 @@ module OpenAI ?request_options: OpenAI::request_opts ) -> OpenAI::Models::Audio::transcription_create_response - def stream_raw: ( + def create_streaming: ( file: IO | StringIO, model: OpenAI::Models::Audio::TranscriptionCreateParams::model, ?include: ::Array[OpenAI::Models::Audio::transcription_include], diff --git a/sig/openai/resources/beta/threads/runs.rbs b/sig/openai/resources/beta/threads/runs.rbs index 5a1bb9fd..df8265f2 100644 --- a/sig/openai/resources/beta/threads/runs.rbs +++ b/sig/openai/resources/beta/threads/runs.rbs @@ -27,7 +27,7 @@ module OpenAI ?request_options: OpenAI::request_opts ) -> OpenAI::Models::Beta::Threads::Run - def stream_raw: ( + def create_stream_raw: ( String thread_id, assistant_id: String, ?include: ::Array[OpenAI::Models::Beta::Threads::Runs::run_step_include], @@ -84,7 +84,7 @@ module OpenAI ?request_options: OpenAI::request_opts ) -> OpenAI::Models::Beta::Threads::Run - def submit_tool_stream_raw: ( + def submit_tool_outputs_stream_raw: ( String run_id, thread_id: String, tool_outputs: ::Array[OpenAI::Models::Beta::Threads::RunSubmitToolOutputsParams::ToolOutput], diff --git a/sig/openai/resources/completions.rbs b/sig/openai/resources/completions.rbs index b786a2e1..c266af82 100644 --- a/sig/openai/resources/completions.rbs +++ b/sig/openai/resources/completions.rbs @@ -22,7 +22,7 @@ module OpenAI ?request_options: OpenAI::request_opts ) -> OpenAI::Models::Completion - def stream_raw: ( + def create_streaming: ( model: OpenAI::Models::CompletionCreateParams::model, prompt: OpenAI::Models::CompletionCreateParams::prompt?, ?best_of: Integer?,