Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions .stats.yml
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
configured_endpoints: 95
openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/openai%2Fopenai-32de3bc513663c5fac922c49be41c222b6ee8c0b841d8966bcdfa489d441daa3.yml
openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/openai%2Fopenai-44b20fa9d24544217fe6bb48852037537030a1ad29b202936425110744fe66fb.yml
openapi_spec_hash: ea86343b5e9858a74e85da8ab2c532f6
config_hash: 43dc8df20ffec9d1503f91866cb2b7d9
config_hash: 69e3afd56ccb0f0f822a7a9dc130fc99
2 changes: 1 addition & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -73,7 +73,7 @@ We provide support for streaming responses using Server Side Events (SSE).

```ruby
stream =
openai.chat.completions.create_streaming(
openai.chat.completions.stream_raw(
messages: [{
role: "user",
content: "Say this is a test"
Expand Down
2 changes: 1 addition & 1 deletion lib/openai/models/beta/thread_create_and_run_params.rb
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ module Models
module Beta
# @see OpenAI::Resources::Beta::Threads#create_and_run
#
# @see OpenAI::Resources::Beta::Threads#create_and_run_streaming
# @see OpenAI::Resources::Beta::Threads#stream_raw
class ThreadCreateAndRunParams < OpenAI::Internal::Type::BaseModel
# @!parse
# extend OpenAI::Internal::Type::RequestParameters::Converter
Expand Down
2 changes: 1 addition & 1 deletion lib/openai/models/beta/threads/run.rb
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ module Beta
module Threads
# @see OpenAI::Resources::Beta::Threads::Runs#create
#
# @see OpenAI::Resources::Beta::Threads::Runs#create_streaming
# @see OpenAI::Resources::Beta::Threads::Runs#create_stream_raw
class Run < OpenAI::Internal::Type::BaseModel
# @!attribute id
# The identifier, which can be referenced in API endpoints.
Expand Down
2 changes: 1 addition & 1 deletion lib/openai/models/beta/threads/run_create_params.rb
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ module Beta
module Threads
# @see OpenAI::Resources::Beta::Threads::Runs#create
#
# @see OpenAI::Resources::Beta::Threads::Runs#create_streaming
# @see OpenAI::Resources::Beta::Threads::Runs#create_stream_raw
class RunCreateParams < OpenAI::Internal::Type::BaseModel
# @!parse
# extend OpenAI::Internal::Type::RequestParameters::Converter
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ module Beta
module Threads
# @see OpenAI::Resources::Beta::Threads::Runs#submit_tool_outputs
#
# @see OpenAI::Resources::Beta::Threads::Runs#submit_tool_outputs_streaming
# @see OpenAI::Resources::Beta::Threads::Runs#submit_tool_outputs_stream_raw
class RunSubmitToolOutputsParams < OpenAI::Internal::Type::BaseModel
# @!parse
# extend OpenAI::Internal::Type::RequestParameters::Converter
Expand Down
2 changes: 1 addition & 1 deletion lib/openai/models/chat/chat_completion.rb
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ module Models
module Chat
# @see OpenAI::Resources::Chat::Completions#create
#
# @see OpenAI::Resources::Chat::Completions#create_streaming
# @see OpenAI::Resources::Chat::Completions#stream_raw
class ChatCompletion < OpenAI::Internal::Type::BaseModel
# @!attribute id
# A unique identifier for the chat completion.
Expand Down
2 changes: 1 addition & 1 deletion lib/openai/models/chat/completion_create_params.rb
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ module Models
module Chat
# @see OpenAI::Resources::Chat::Completions#create
#
# @see OpenAI::Resources::Chat::Completions#create_streaming
# @see OpenAI::Resources::Chat::Completions#stream_raw
class CompletionCreateParams < OpenAI::Internal::Type::BaseModel
# @!parse
# extend OpenAI::Internal::Type::RequestParameters::Converter
Expand Down
2 changes: 1 addition & 1 deletion lib/openai/models/responses/response.rb
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ module Models
module Responses
# @see OpenAI::Resources::Responses#create
#
# @see OpenAI::Resources::Responses#create_streaming
# @see OpenAI::Resources::Responses#stream_raw
class Response < OpenAI::Internal::Type::BaseModel
# @!attribute id
# Unique identifier for this Response.
Expand Down
2 changes: 1 addition & 1 deletion lib/openai/models/responses/response_create_params.rb
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ module Models
module Responses
# @see OpenAI::Resources::Responses#create
#
# @see OpenAI::Resources::Responses#create_streaming
# @see OpenAI::Resources::Responses#stream_raw
class ResponseCreateParams < OpenAI::Internal::Type::BaseModel
# @!parse
# extend OpenAI::Internal::Type::RequestParameters::Converter
Expand Down
9 changes: 4 additions & 5 deletions lib/openai/resources/beta/threads.rb
Original file line number Diff line number Diff line change
Expand Up @@ -94,8 +94,7 @@ def delete(thread_id, params = {})
)
end

# See {OpenAI::Resources::Beta::Threads#create_and_run_streaming} for streaming
# counterpart.
# See {OpenAI::Resources::Beta::Threads#stream_raw} for streaming counterpart.
#
# Create a thread and run it in one request.
#
Expand Down Expand Up @@ -124,7 +123,7 @@ def delete(thread_id, params = {})
def create_and_run(params)
parsed, options = OpenAI::Models::Beta::ThreadCreateAndRunParams.dump_request(params)
if parsed[:stream]
message = "Please use `#create_and_run_streaming` for the streaming use case."
message = "Please use `#stream_raw` for the streaming use case."
raise ArgumentError.new(message)
end
@client.request(
Expand All @@ -141,7 +140,7 @@ def create_and_run(params)
#
# Create a thread and run it in one request.
#
# @overload create_and_run_streaming(assistant_id:, instructions: nil, max_completion_tokens: nil, max_prompt_tokens: nil, metadata: nil, model: nil, parallel_tool_calls: nil, response_format: nil, temperature: nil, thread: nil, tool_choice: nil, tool_resources: nil, tools: nil, top_p: nil, truncation_strategy: nil, request_options: {})
# @overload stream_raw(assistant_id:, instructions: nil, max_completion_tokens: nil, max_prompt_tokens: nil, metadata: nil, model: nil, parallel_tool_calls: nil, response_format: nil, temperature: nil, thread: nil, tool_choice: nil, tool_resources: nil, tools: nil, top_p: nil, truncation_strategy: nil, request_options: {})
#
# @param assistant_id [String]
# @param instructions [String, nil]
Expand All @@ -163,7 +162,7 @@ def create_and_run(params)
# @return [OpenAI::Internal::Stream<OpenAI::Models::Beta::AssistantStreamEvent::ThreadCreated, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunCreated, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunQueued, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunInProgress, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunRequiresAction, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunCompleted, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunIncomplete, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunFailed, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunCancelling, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunCancelled, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunExpired, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunStepCreated, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunStepInProgress, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunStepDelta, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunStepCompleted, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunStepFailed, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunStepCancelled, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunStepExpired, OpenAI::Models::Beta::AssistantStreamEvent::ThreadMessageCreated, OpenAI::Models::Beta::AssistantStreamEvent::ThreadMessageInProgress, OpenAI::Models::Beta::AssistantStreamEvent::ThreadMessageDelta, OpenAI::Models::Beta::AssistantStreamEvent::ThreadMessageCompleted, OpenAI::Models::Beta::AssistantStreamEvent::ThreadMessageIncomplete, OpenAI::Models::Beta::AssistantStreamEvent::ErrorEvent>]
#
# @see OpenAI::Models::Beta::ThreadCreateAndRunParams
def create_and_run_streaming(params)
def stream_raw(params)
parsed, options = OpenAI::Models::Beta::ThreadCreateAndRunParams.dump_request(params)
unless parsed.fetch(:stream, true)
message = "Please use `#create_and_run` for the non-streaming use case."
Expand Down
16 changes: 8 additions & 8 deletions lib/openai/resources/beta/threads/runs.rb
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ class Runs
# @return [OpenAI::Resources::Beta::Threads::Runs::Steps]
attr_reader :steps

# See {OpenAI::Resources::Beta::Threads::Runs#create_streaming} for streaming
# See {OpenAI::Resources::Beta::Threads::Runs#create_stream_raw} for streaming
# counterpart.
#
# Create a run.
Expand Down Expand Up @@ -41,7 +41,7 @@ class Runs
def create(thread_id, params)
parsed, options = OpenAI::Models::Beta::Threads::RunCreateParams.dump_request(params)
if parsed[:stream]
message = "Please use `#create_streaming` for the streaming use case."
message = "Please use `#create_stream_raw` for the streaming use case."
raise ArgumentError.new(message)
end
query_params = [:include]
Expand All @@ -60,7 +60,7 @@ def create(thread_id, params)
#
# Create a run.
#
# @overload create_streaming(thread_id, assistant_id:, include: nil, additional_instructions: nil, additional_messages: nil, instructions: nil, max_completion_tokens: nil, max_prompt_tokens: nil, metadata: nil, model: nil, parallel_tool_calls: nil, reasoning_effort: nil, response_format: nil, temperature: nil, tool_choice: nil, tools: nil, top_p: nil, truncation_strategy: nil, request_options: {})
# @overload create_stream_raw(thread_id, assistant_id:, include: nil, additional_instructions: nil, additional_messages: nil, instructions: nil, max_completion_tokens: nil, max_prompt_tokens: nil, metadata: nil, model: nil, parallel_tool_calls: nil, reasoning_effort: nil, response_format: nil, temperature: nil, tool_choice: nil, tools: nil, top_p: nil, truncation_strategy: nil, request_options: {})
#
# @param thread_id [String]
# @param assistant_id [String]
Expand All @@ -85,7 +85,7 @@ def create(thread_id, params)
# @return [OpenAI::Internal::Stream<OpenAI::Models::Beta::AssistantStreamEvent::ThreadCreated, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunCreated, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunQueued, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunInProgress, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunRequiresAction, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunCompleted, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunIncomplete, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunFailed, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunCancelling, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunCancelled, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunExpired, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunStepCreated, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunStepInProgress, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunStepDelta, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunStepCompleted, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunStepFailed, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunStepCancelled, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunStepExpired, OpenAI::Models::Beta::AssistantStreamEvent::ThreadMessageCreated, OpenAI::Models::Beta::AssistantStreamEvent::ThreadMessageInProgress, OpenAI::Models::Beta::AssistantStreamEvent::ThreadMessageDelta, OpenAI::Models::Beta::AssistantStreamEvent::ThreadMessageCompleted, OpenAI::Models::Beta::AssistantStreamEvent::ThreadMessageIncomplete, OpenAI::Models::Beta::AssistantStreamEvent::ErrorEvent>]
#
# @see OpenAI::Models::Beta::Threads::RunCreateParams
def create_streaming(thread_id, params)
def create_stream_raw(thread_id, params)
parsed, options = OpenAI::Models::Beta::Threads::RunCreateParams.dump_request(params)
unless parsed.fetch(:stream, true)
message = "Please use `#create` for the non-streaming use case."
Expand Down Expand Up @@ -208,7 +208,7 @@ def cancel(run_id, params)
)
end

# See {OpenAI::Resources::Beta::Threads::Runs#submit_tool_outputs_streaming} for
# See {OpenAI::Resources::Beta::Threads::Runs#submit_tool_outputs_stream_raw} for
# streaming counterpart.
#
# When a run has the `status: "requires_action"` and `required_action.type` is
Expand All @@ -229,7 +229,7 @@ def cancel(run_id, params)
def submit_tool_outputs(run_id, params)
parsed, options = OpenAI::Models::Beta::Threads::RunSubmitToolOutputsParams.dump_request(params)
if parsed[:stream]
message = "Please use `#submit_tool_outputs_streaming` for the streaming use case."
message = "Please use `#submit_tool_outputs_stream_raw` for the streaming use case."
raise ArgumentError.new(message)
end
thread_id =
Expand All @@ -253,7 +253,7 @@ def submit_tool_outputs(run_id, params)
# tool calls once they're all completed. All outputs must be submitted in a single
# request.
#
# @overload submit_tool_outputs_streaming(run_id, thread_id:, tool_outputs:, request_options: {})
# @overload submit_tool_outputs_stream_raw(run_id, thread_id:, tool_outputs:, request_options: {})
#
# @param run_id [String]
# @param thread_id [String]
Expand All @@ -263,7 +263,7 @@ def submit_tool_outputs(run_id, params)
# @return [OpenAI::Internal::Stream<OpenAI::Models::Beta::AssistantStreamEvent::ThreadCreated, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunCreated, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunQueued, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunInProgress, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunRequiresAction, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunCompleted, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunIncomplete, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunFailed, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunCancelling, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunCancelled, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunExpired, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunStepCreated, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunStepInProgress, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunStepDelta, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunStepCompleted, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunStepFailed, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunStepCancelled, OpenAI::Models::Beta::AssistantStreamEvent::ThreadRunStepExpired, OpenAI::Models::Beta::AssistantStreamEvent::ThreadMessageCreated, OpenAI::Models::Beta::AssistantStreamEvent::ThreadMessageInProgress, OpenAI::Models::Beta::AssistantStreamEvent::ThreadMessageDelta, OpenAI::Models::Beta::AssistantStreamEvent::ThreadMessageCompleted, OpenAI::Models::Beta::AssistantStreamEvent::ThreadMessageIncomplete, OpenAI::Models::Beta::AssistantStreamEvent::ErrorEvent>]
#
# @see OpenAI::Models::Beta::Threads::RunSubmitToolOutputsParams
def submit_tool_outputs_streaming(run_id, params)
def submit_tool_outputs_stream_raw(run_id, params)
parsed, options = OpenAI::Models::Beta::Threads::RunSubmitToolOutputsParams.dump_request(params)
unless parsed.fetch(:stream, true)
message = "Please use `#submit_tool_outputs` for the non-streaming use case."
Expand Down
9 changes: 4 additions & 5 deletions lib/openai/resources/chat/completions.rb
Original file line number Diff line number Diff line change
Expand Up @@ -7,8 +7,7 @@ class Completions
# @return [OpenAI::Resources::Chat::Completions::Messages]
attr_reader :messages

# See {OpenAI::Resources::Chat::Completions#create_streaming} for streaming
# counterpart.
# See {OpenAI::Resources::Chat::Completions#stream_raw} for streaming counterpart.
#
# **Starting a new project?** We recommend trying
# [Responses](https://platform.openai.com/docs/api-reference/responses) to take
Expand Down Expand Up @@ -68,7 +67,7 @@ class Completions
def create(params)
parsed, options = OpenAI::Models::Chat::CompletionCreateParams.dump_request(params)
if parsed[:stream]
message = "Please use `#create_streaming` for the streaming use case."
message = "Please use `#stream_raw` for the streaming use case."
raise ArgumentError.new(message)
end
@client.request(
Expand Down Expand Up @@ -100,7 +99,7 @@ def create(params)
# unsupported parameters in reasoning models,
# [refer to the reasoning guide](https://platform.openai.com/docs/guides/reasoning).
#
# @overload create_streaming(messages:, model:, audio: nil, frequency_penalty: nil, function_call: nil, functions: nil, logit_bias: nil, logprobs: nil, max_completion_tokens: nil, max_tokens: nil, metadata: nil, modalities: nil, n: nil, parallel_tool_calls: nil, prediction: nil, presence_penalty: nil, reasoning_effort: nil, response_format: nil, seed: nil, service_tier: nil, stop: nil, store: nil, stream_options: nil, temperature: nil, tool_choice: nil, tools: nil, top_logprobs: nil, top_p: nil, user: nil, web_search_options: nil, request_options: {})
# @overload stream_raw(messages:, model:, audio: nil, frequency_penalty: nil, function_call: nil, functions: nil, logit_bias: nil, logprobs: nil, max_completion_tokens: nil, max_tokens: nil, metadata: nil, modalities: nil, n: nil, parallel_tool_calls: nil, prediction: nil, presence_penalty: nil, reasoning_effort: nil, response_format: nil, seed: nil, service_tier: nil, stop: nil, store: nil, stream_options: nil, temperature: nil, tool_choice: nil, tools: nil, top_logprobs: nil, top_p: nil, user: nil, web_search_options: nil, request_options: {})
#
# @param messages [Array<OpenAI::Models::Chat::ChatCompletionDeveloperMessageParam, OpenAI::Models::Chat::ChatCompletionSystemMessageParam, OpenAI::Models::Chat::ChatCompletionUserMessageParam, OpenAI::Models::Chat::ChatCompletionAssistantMessageParam, OpenAI::Models::Chat::ChatCompletionToolMessageParam, OpenAI::Models::Chat::ChatCompletionFunctionMessageParam>]
# @param model [String, Symbol, OpenAI::Models::ChatModel]
Expand Down Expand Up @@ -137,7 +136,7 @@ def create(params)
# @return [OpenAI::Internal::Stream<OpenAI::Models::Chat::ChatCompletionChunk>]
#
# @see OpenAI::Models::Chat::CompletionCreateParams
def create_streaming(params)
def stream_raw(params)
parsed, options = OpenAI::Models::Chat::CompletionCreateParams.dump_request(params)
unless parsed.fetch(:stream, true)
message = "Please use `#create` for the non-streaming use case."
Expand Down
Loading