diff --git a/lib/openai.rb b/lib/openai.rb index 8e87e568..9b3ac42c 100644 --- a/lib/openai.rb +++ b/lib/openai.rb @@ -48,7 +48,7 @@ require_relative "openai/internal/type/base_page" require_relative "openai/internal/type/base_stream" require_relative "openai/internal/type/request_parameters" -require_relative "openai/aliases" +require_relative "openai/internal" require_relative "openai/request_options" require_relative "openai/errors" require_relative "openai/internal/transport/base_client" diff --git a/lib/openai/aliases.rb b/lib/openai/aliases.rb deleted file mode 100644 index da013a3e..00000000 --- a/lib/openai/aliases.rb +++ /dev/null @@ -1,19 +0,0 @@ -# frozen_string_literal: true - -module OpenAI - Unknown = OpenAI::Internal::Type::Unknown - - BooleanModel = OpenAI::Internal::Type::BooleanModel - - Enum = OpenAI::Internal::Type::Enum - - Union = OpenAI::Internal::Type::Union - - ArrayOf = OpenAI::Internal::Type::ArrayOf - - HashOf = OpenAI::Internal::Type::HashOf - - BaseModel = OpenAI::Internal::Type::BaseModel - - RequestParameters = OpenAI::Internal::Type::RequestParameters -end diff --git a/lib/openai/errors.rb b/lib/openai/errors.rb index 4b5fc99f..73e74f52 100644 --- a/lib/openai/errors.rb +++ b/lib/openai/errors.rb @@ -225,32 +225,4 @@ class InternalServerError < OpenAI::Errors::APIStatusError HTTP_STATUS = (500..) end end - - Error = OpenAI::Errors::Error - - ConversionError = OpenAI::Errors::ConversionError - - APIError = OpenAI::Errors::APIError - - APIStatusError = OpenAI::Errors::APIStatusError - - APIConnectionError = OpenAI::Errors::APIConnectionError - - APITimeoutError = OpenAI::Errors::APITimeoutError - - BadRequestError = OpenAI::Errors::BadRequestError - - AuthenticationError = OpenAI::Errors::AuthenticationError - - PermissionDeniedError = OpenAI::Errors::PermissionDeniedError - - NotFoundError = OpenAI::Errors::NotFoundError - - ConflictError = OpenAI::Errors::ConflictError - - UnprocessableEntityError = OpenAI::Errors::UnprocessableEntityError - - RateLimitError = OpenAI::Errors::RateLimitError - - InternalServerError = OpenAI::Errors::InternalServerError end diff --git a/lib/openai/internal.rb b/lib/openai/internal.rb new file mode 100644 index 00000000..42050499 --- /dev/null +++ b/lib/openai/internal.rb @@ -0,0 +1,8 @@ +# frozen_string_literal: true + +module OpenAI + # @api private + module Internal + OMIT = Object.new.freeze + end +end diff --git a/lib/openai/internal/transport/base_client.rb b/lib/openai/internal/transport/base_client.rb index 013e2d64..cd7979be 100644 --- a/lib/openai/internal/transport/base_client.rb +++ b/lib/openai/internal/transport/base_client.rb @@ -339,7 +339,7 @@ def initialize( begin status, response, stream = @requester.execute(input) - rescue OpenAI::APIConnectionError => e + rescue OpenAI::Errors::APIConnectionError => e status = e end @@ -361,7 +361,7 @@ def initialize( retry_count: retry_count, send_retry_header: send_retry_header ) - in OpenAI::APIConnectionError if retry_count >= max_retries + in OpenAI::Errors::APIConnectionError if retry_count >= max_retries raise status in (400..) if retry_count >= max_retries || !self.class.should_retry?(status, headers: response) decoded = Kernel.then do @@ -421,7 +421,7 @@ def initialize( # @return [Object] def request(req) self.class.validate!(req) - model = req.fetch(:model) { OpenAI::Unknown } + model = req.fetch(:model) { OpenAI::Internal::Type::Unknown } opts = req[:options].to_h OpenAI::RequestOptions.validate!(opts) request = build_request(req.except(:options), opts) diff --git a/lib/openai/internal/type/array_of.rb b/lib/openai/internal/type/array_of.rb index 65fa4ccb..5c67fc15 100644 --- a/lib/openai/internal/type/array_of.rb +++ b/lib/openai/internal/type/array_of.rb @@ -33,7 +33,9 @@ def ===(other) = other.is_a?(Array) && other.all?(item_type) # # @return [Boolean] def ==(other) - other.is_a?(OpenAI::ArrayOf) && other.nilable? == nilable? && other.item_type == item_type + # rubocop:disable Layout/LineLength + other.is_a?(OpenAI::Internal::Type::ArrayOf) && other.nilable? == nilable? && other.item_type == item_type + # rubocop:enable Layout/LineLength end # @api private diff --git a/lib/openai/internal/type/base_model.rb b/lib/openai/internal/type/base_model.rb index 8984ef64..5dde350e 100644 --- a/lib/openai/internal/type/base_model.rb +++ b/lib/openai/internal/type/base_model.rb @@ -23,7 +23,7 @@ class << self # # @return [Hash{Symbol=>Hash{Symbol=>Object}}] def known_fields - @known_fields ||= (self < OpenAI::BaseModel ? superclass.known_fields.dup : {}) + @known_fields ||= (self < OpenAI::Internal::Type::BaseModel ? superclass.known_fields.dup : {}) end # @api private @@ -64,14 +64,7 @@ def fields setter = "#{name_sym}=" api_name = info.fetch(:api_name, name_sym) nilable = info[:nil?] - const = if required && !nilable - info.fetch( - :const, - OpenAI::Internal::Util::OMIT - ) - else - OpenAI::Internal::Util::OMIT - end + const = required && !nilable ? info.fetch(:const, OpenAI::Internal::OMIT) : OpenAI::Internal::OMIT [name_sym, setter].each { undef_method(_1) } if known_fields.key?(name_sym) @@ -89,7 +82,7 @@ def fields define_method(name_sym) do target = type_fn.call - value = @data.fetch(name_sym) { const == OpenAI::Internal::Util::OMIT ? nil : const } + value = @data.fetch(name_sym) { const == OpenAI::Internal::OMIT ? nil : const } state = {strictness: :strong, exactness: {yes: 0, no: 0, maybe: 0}, branched: 0} if (nilable || !required) && value.nil? nil @@ -105,7 +98,7 @@ def fields # rubocop:disable Layout/LineLength message = "Failed to parse #{cls}.#{__method__} from #{value.class} to #{target.inspect}. To get the unparsed API response, use #{cls}[:#{__method__}]." # rubocop:enable Layout/LineLength - raise OpenAI::ConversionError.new(message) + raise OpenAI::Errors::ConversionError.new(message) end end @@ -175,7 +168,7 @@ def optional(name_sym, type_info, spec = {}) # @param other [Object] # # @return [Boolean] - def ==(other) = other.is_a?(Class) && other <= OpenAI::BaseModel && other.fields == fields + def ==(other) = other.is_a?(Class) && other <= OpenAI::Internal::Type::BaseModel && other.fields == fields end # @param other [Object] @@ -186,7 +179,7 @@ def ==(other) = self.class == other.class && @data == other.to_h class << self # @api private # - # @param value [OpenAI::BaseModel, Hash{Object=>Object}, Object] + # @param value [OpenAI::Internal::Type::BaseModel, Hash{Object=>Object}, Object] # # @param state [Hash{Symbol=>Object}] . # @@ -196,7 +189,7 @@ class << self # # @option state [Integer] :branched # - # @return [OpenAI::BaseModel, Object] + # @return [OpenAI::Internal::Type::BaseModel, Object] def coerce(value, state:) exactness = state.fetch(:exactness) @@ -221,7 +214,7 @@ def coerce(value, state:) api_name, nilable, const = field.fetch_values(:api_name, :nilable, :const) unless val.key?(api_name) - if required && mode != :dump && const == OpenAI::Internal::Util::OMIT + if required && mode != :dump && const == OpenAI::Internal::OMIT exactness[nilable ? :maybe : :no] += 1 else exactness[:yes] += 1 @@ -255,7 +248,7 @@ def coerce(value, state:) # @api private # - # @param value [OpenAI::BaseModel, Object] + # @param value [OpenAI::Internal::Type::BaseModel, Object] # # @return [Hash{Object=>Object}, Object] def dump(value) @@ -284,7 +277,7 @@ def dump(value) known_fields.each_value do |field| mode, api_name, const = field.fetch_values(:mode, :api_name, :const) - next if mode == :coerce || acc.key?(api_name) || const == OpenAI::Internal::Util::OMIT + next if mode == :coerce || acc.key?(api_name) || const == OpenAI::Internal::OMIT acc.store(api_name, const) end @@ -351,13 +344,13 @@ def to_yaml(*a) = self.class.dump(self).to_yaml(*a) # Create a new instance of a model. # - # @param data [Hash{Symbol=>Object}, OpenAI::BaseModel] + # @param data [Hash{Symbol=>Object}, OpenAI::Internal::Type::BaseModel] def initialize(data = {}) case OpenAI::Internal::Util.coerce_hash(data) in Hash => coerced @data = coerced else - raise ArgumentError.new("Expected a #{Hash} or #{OpenAI::BaseModel}, got #{data.inspect}") + raise ArgumentError.new("Expected a #{Hash} or #{OpenAI::Internal::Type::BaseModel}, got #{data.inspect}") end end diff --git a/lib/openai/internal/type/boolean_model.rb b/lib/openai/internal/type/boolean_model.rb index 369fd741..cf1dcf00 100644 --- a/lib/openai/internal/type/boolean_model.rb +++ b/lib/openai/internal/type/boolean_model.rb @@ -19,7 +19,7 @@ def self.===(other) = other == true || other == false # @param other [Object] # # @return [Boolean] - def self.==(other) = other.is_a?(Class) && other <= OpenAI::BooleanModel + def self.==(other) = other.is_a?(Class) && other <= OpenAI::Internal::Type::BooleanModel class << self # @api private diff --git a/lib/openai/internal/type/converter.rb b/lib/openai/internal/type/converter.rb index 29aad06d..ee4589cf 100644 --- a/lib/openai/internal/type/converter.rb +++ b/lib/openai/internal/type/converter.rb @@ -31,10 +31,10 @@ def coerce(value, state:) = (raise NotImplementedError) def dump(value) case value in Array - value.map { OpenAI::Unknown.dump(_1) } + value.map { OpenAI::Internal::Type::Unknown.dump(_1) } in Hash - value.transform_values { OpenAI::Unknown.dump(_1) } - in OpenAI::BaseModel + value.transform_values { OpenAI::Internal::Type::Unknown.dump(_1) } + in OpenAI::Internal::Type::BaseModel value.class.dump(value) else value @@ -64,7 +64,7 @@ def type_info(spec) in Hash type_info(spec.slice(:const, :enum, :union).first&.last) in true | false - -> { OpenAI::BooleanModel } + -> { OpenAI::Internal::Type::BooleanModel } in OpenAI::Internal::Type::Converter | Class | Symbol -> { spec } in NilClass | Integer | Float @@ -209,7 +209,7 @@ def coerce( # # @return [Object] def dump(target, value) - target.is_a?(OpenAI::Internal::Type::Converter) ? target.dump(value) : OpenAI::Unknown.dump(value) + target.is_a?(OpenAI::Internal::Type::Converter) ? target.dump(value) : OpenAI::Internal::Type::Unknown.dump(value) end end end diff --git a/lib/openai/internal/type/enum.rb b/lib/openai/internal/type/enum.rb index 861f0ac9..de4e7a7a 100644 --- a/lib/openai/internal/type/enum.rb +++ b/lib/openai/internal/type/enum.rb @@ -62,7 +62,7 @@ def ===(other) = values.include?(other) # # @return [Boolean] def ==(other) - other.is_a?(Module) && other.singleton_class <= OpenAI::Enum && other.values.to_set == values.to_set + other.is_a?(Module) && other.singleton_class <= OpenAI::Internal::Type::Enum && other.values.to_set == values.to_set end # @api private diff --git a/lib/openai/internal/type/hash_of.rb b/lib/openai/internal/type/hash_of.rb index d07f7ccd..eb5b546e 100644 --- a/lib/openai/internal/type/hash_of.rb +++ b/lib/openai/internal/type/hash_of.rb @@ -48,7 +48,9 @@ def ===(other) # # @return [Boolean] def ==(other) - other.is_a?(OpenAI::HashOf) && other.nilable? == nilable? && other.item_type == item_type + # rubocop:disable Layout/LineLength + other.is_a?(OpenAI::Internal::Type::HashOf) && other.nilable? == nilable? && other.item_type == item_type + # rubocop:enable Layout/LineLength end # @api private diff --git a/lib/openai/internal/type/request_parameters.rb b/lib/openai/internal/type/request_parameters.rb index 8a693e53..93ebb068 100644 --- a/lib/openai/internal/type/request_parameters.rb +++ b/lib/openai/internal/type/request_parameters.rb @@ -12,7 +12,7 @@ module RequestParameters # @param mod [Module] def self.included(mod) - return unless mod <= OpenAI::BaseModel + return unless mod <= OpenAI::Internal::Type::BaseModel mod.extend(OpenAI::Internal::Type::RequestParameters::Converter) mod.optional(:request_options, OpenAI::RequestOptions) diff --git a/lib/openai/internal/type/union.rb b/lib/openai/internal/type/union.rb index bf2b376e..18c6cf0e 100644 --- a/lib/openai/internal/type/union.rb +++ b/lib/openai/internal/type/union.rb @@ -93,14 +93,14 @@ def variants = derefed_variants.map(&:last) # @return [OpenAI::Internal::Type::Converter, Class, nil] private def resolve_variant(value) case [@discriminator, value] - in [_, OpenAI::BaseModel] + in [_, OpenAI::Internal::Type::BaseModel] value.class in [Symbol, Hash] key = value.fetch(@discriminator) do - value.fetch(@discriminator.to_s, OpenAI::Internal::Util::OMIT) + value.fetch(@discriminator.to_s, OpenAI::Internal::OMIT) end - return nil if key == OpenAI::Internal::Util::OMIT + return nil if key == OpenAI::Internal::OMIT key = key.to_sym if key.is_a?(String) known_variants.find { |k,| k == key }&.last&.call @@ -126,7 +126,7 @@ def ===(other) # @return [Boolean] def ==(other) # rubocop:disable Layout/LineLength - other.is_a?(Module) && other.singleton_class <= OpenAI::Union && other.derefed_variants == derefed_variants + other.is_a?(Module) && other.singleton_class <= OpenAI::Internal::Type::Union && other.derefed_variants == derefed_variants # rubocop:enable Layout/LineLength end diff --git a/lib/openai/internal/type/unknown.rb b/lib/openai/internal/type/unknown.rb index 589bd2ba..5d06c449 100644 --- a/lib/openai/internal/type/unknown.rb +++ b/lib/openai/internal/type/unknown.rb @@ -21,7 +21,7 @@ def self.===(other) = true # @param other [Object] # # @return [Boolean] - def self.==(other) = other.is_a?(Class) && other <= OpenAI::Unknown + def self.==(other) = other.is_a?(Class) && other <= OpenAI::Internal::Type::Unknown class << self # @api private diff --git a/lib/openai/internal/util.rb b/lib/openai/internal/util.rb index 6fb201e3..da915e57 100644 --- a/lib/openai/internal/util.rb +++ b/lib/openai/internal/util.rb @@ -148,13 +148,6 @@ def coerce_hash(input) end end - # Use this to indicate that a value should be explicitly removed from a data - # structure when using `OpenAI::Internal::Util.deep_merge`. - # - # e.g. merging `{a: 1}` and `{a: OMIT}` should produce `{}`, where merging - # `{a: 1}` and `{}` would produce `{a: 1}`. - OMIT = Object.new.freeze - class << self # @api private # @@ -166,9 +159,9 @@ class << self private def deep_merge_lr(lhs, rhs, concat: false) case [lhs, rhs, concat] in [Hash, Hash, _] - rhs_cleaned = rhs.reject { _2 == OpenAI::Internal::Util::OMIT } + rhs_cleaned = rhs.reject { _2 == OpenAI::Internal::OMIT } lhs - .reject { |key, _| rhs[key] == OpenAI::Internal::Util::OMIT } + .reject { |key, _| rhs[key] == OpenAI::Internal::OMIT } .merge(rhs_cleaned) do |_, old_val, new_val| deep_merge_lr(old_val, new_val, concat: concat) end diff --git a/lib/openai/models/all_models.rb b/lib/openai/models/all_models.rb index f09c1abd..6ea63fb0 100644 --- a/lib/openai/models/all_models.rb +++ b/lib/openai/models/all_models.rb @@ -3,7 +3,7 @@ module OpenAI module Models module AllModels - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String diff --git a/lib/openai/models/audio/speech_create_params.rb b/lib/openai/models/audio/speech_create_params.rb index 0729f543..94b31a52 100644 --- a/lib/openai/models/audio/speech_create_params.rb +++ b/lib/openai/models/audio/speech_create_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Audio # @see OpenAI::Resources::Audio::Speech#create - class SpeechCreateParams < OpenAI::BaseModel + class SpeechCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -75,12 +75,12 @@ class SpeechCreateParams < OpenAI::BaseModel # # # def initialize(input:, model:, voice:, instructions: nil, response_format: nil, speed: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # One of the available [TTS models](https://platform.openai.com/docs/models#tts): # `tts-1`, `tts-1-hd` or `gpt-4o-mini-tts`. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String @@ -97,7 +97,7 @@ module Model # `verse`. Previews of the voices are available in the # [Text to speech guide](https://platform.openai.com/docs/guides/text-to-speech#voice-options). module Voice - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String @@ -147,7 +147,7 @@ module Voice # The format to audio in. Supported formats are `mp3`, `opus`, `aac`, `flac`, # `wav`, and `pcm`. module ResponseFormat - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum MP3 = :mp3 OPUS = :opus diff --git a/lib/openai/models/audio/speech_model.rb b/lib/openai/models/audio/speech_model.rb index fad7aa94..504b7319 100644 --- a/lib/openai/models/audio/speech_model.rb +++ b/lib/openai/models/audio/speech_model.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Audio module SpeechModel - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TTS_1 = :"tts-1" TTS_1_HD = :"tts-1-hd" diff --git a/lib/openai/models/audio/transcription.rb b/lib/openai/models/audio/transcription.rb index 8185ea6c..abe567ce 100644 --- a/lib/openai/models/audio/transcription.rb +++ b/lib/openai/models/audio/transcription.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Audio - class Transcription < OpenAI::BaseModel + class Transcription < OpenAI::Internal::Type::BaseModel # @!attribute text # The transcribed text. # @@ -16,7 +16,7 @@ class Transcription < OpenAI::BaseModel # to the `include` array. # # @return [Array, nil] - optional :logprobs, -> { OpenAI::ArrayOf[OpenAI::Models::Audio::Transcription::Logprob] } + optional :logprobs, -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Audio::Transcription::Logprob] } # @!parse # # @return [Array] @@ -31,9 +31,9 @@ class Transcription < OpenAI::BaseModel # # # def initialize(text:, logprobs: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class Logprob < OpenAI::BaseModel + class Logprob < OpenAI::Internal::Type::BaseModel # @!attribute [r] token # The token in the transcription. # @@ -48,7 +48,7 @@ class Logprob < OpenAI::BaseModel # The bytes of the token. # # @return [Array, nil] - optional :bytes, OpenAI::ArrayOf[Float] + optional :bytes, OpenAI::Internal::Type::ArrayOf[Float] # @!parse # # @return [Array] @@ -71,7 +71,7 @@ class Logprob < OpenAI::BaseModel # # # def initialize(token: nil, bytes: nil, logprob: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/audio/transcription_create_params.rb b/lib/openai/models/audio/transcription_create_params.rb index 063af4b9..f3dc4d0d 100644 --- a/lib/openai/models/audio/transcription_create_params.rb +++ b/lib/openai/models/audio/transcription_create_params.rb @@ -6,7 +6,7 @@ module Audio # @see OpenAI::Resources::Audio::Transcriptions#create # # @see OpenAI::Resources::Audio::Transcriptions#stream_raw - class TranscriptionCreateParams < OpenAI::BaseModel + class TranscriptionCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -34,7 +34,8 @@ class TranscriptionCreateParams < OpenAI::BaseModel # `gpt-4o-mini-transcribe`. # # @return [Array, nil] - optional :include, -> { OpenAI::ArrayOf[enum: OpenAI::Models::Audio::TranscriptionInclude] } + optional :include, + -> { OpenAI::Internal::Type::ArrayOf[enum: OpenAI::Models::Audio::TranscriptionInclude] } # @!parse # # @return [Array] @@ -100,7 +101,7 @@ class TranscriptionCreateParams < OpenAI::BaseModel # # @return [Array, nil] optional :timestamp_granularities, - -> { OpenAI::ArrayOf[enum: OpenAI::Models::Audio::TranscriptionCreateParams::TimestampGranularity] } + -> { OpenAI::Internal::Type::ArrayOf[enum: OpenAI::Models::Audio::TranscriptionCreateParams::TimestampGranularity] } # @!parse # # @return [Array] @@ -132,13 +133,13 @@ class TranscriptionCreateParams < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # ID of the model to use. The options are `gpt-4o-transcribe`, # `gpt-4o-mini-transcribe`, and `whisper-1` (which is powered by our open source # Whisper V2 model). module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String @@ -151,7 +152,7 @@ module Model end module TimestampGranularity - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum WORD = :word SEGMENT = :segment diff --git a/lib/openai/models/audio/transcription_create_response.rb b/lib/openai/models/audio/transcription_create_response.rb index efebb7c6..a8026630 100644 --- a/lib/openai/models/audio/transcription_create_response.rb +++ b/lib/openai/models/audio/transcription_create_response.rb @@ -10,7 +10,7 @@ module Audio # # @see OpenAI::Resources::Audio::Transcriptions#stream_raw module TranscriptionCreateResponse - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # Represents a transcription response returned by model, based on the provided input. variant -> { OpenAI::Models::Audio::Transcription } diff --git a/lib/openai/models/audio/transcription_include.rb b/lib/openai/models/audio/transcription_include.rb index 7c76eb59..5cde821b 100644 --- a/lib/openai/models/audio/transcription_include.rb +++ b/lib/openai/models/audio/transcription_include.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Audio module TranscriptionInclude - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum LOGPROBS = :logprobs diff --git a/lib/openai/models/audio/transcription_segment.rb b/lib/openai/models/audio/transcription_segment.rb index b219af36..02990b46 100644 --- a/lib/openai/models/audio/transcription_segment.rb +++ b/lib/openai/models/audio/transcription_segment.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Audio - class TranscriptionSegment < OpenAI::BaseModel + class TranscriptionSegment < OpenAI::Internal::Type::BaseModel # @!attribute id # Unique identifier of the segment. # @@ -65,7 +65,7 @@ class TranscriptionSegment < OpenAI::BaseModel # Array of token IDs for the text content. # # @return [Array] - required :tokens, OpenAI::ArrayOf[Integer] + required :tokens, OpenAI::Internal::Type::ArrayOf[Integer] # @!parse # # @param id [Integer] @@ -95,7 +95,7 @@ class TranscriptionSegment < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/audio/transcription_stream_event.rb b/lib/openai/models/audio/transcription_stream_event.rb index 40437586..171cfd23 100644 --- a/lib/openai/models/audio/transcription_stream_event.rb +++ b/lib/openai/models/audio/transcription_stream_event.rb @@ -8,7 +8,7 @@ module Audio # [create a transcription](https://platform.openai.com/docs/api-reference/audio/create-transcription) # with the `Stream` parameter set to `true`. module TranscriptionStreamEvent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type diff --git a/lib/openai/models/audio/transcription_text_delta_event.rb b/lib/openai/models/audio/transcription_text_delta_event.rb index ec8ca4f4..4a858738 100644 --- a/lib/openai/models/audio/transcription_text_delta_event.rb +++ b/lib/openai/models/audio/transcription_text_delta_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Audio - class TranscriptionTextDeltaEvent < OpenAI::BaseModel + class TranscriptionTextDeltaEvent < OpenAI::Internal::Type::BaseModel # @!attribute delta # The text delta that was additionally transcribed. # @@ -22,7 +22,8 @@ class TranscriptionTextDeltaEvent < OpenAI::BaseModel # with the `include[]` parameter set to `logprobs`. # # @return [Array, nil] - optional :logprobs, -> { OpenAI::ArrayOf[OpenAI::Models::Audio::TranscriptionTextDeltaEvent::Logprob] } + optional :logprobs, + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Audio::TranscriptionTextDeltaEvent::Logprob] } # @!parse # # @return [Array] @@ -40,9 +41,9 @@ class TranscriptionTextDeltaEvent < OpenAI::BaseModel # # # def initialize(delta:, logprobs: nil, type: :"transcript.text.delta", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class Logprob < OpenAI::BaseModel + class Logprob < OpenAI::Internal::Type::BaseModel # @!attribute [r] token # The token that was used to generate the log probability. # @@ -57,7 +58,7 @@ class Logprob < OpenAI::BaseModel # The bytes that were used to generate the log probability. # # @return [Array, nil] - optional :bytes, OpenAI::ArrayOf[OpenAI::Unknown] + optional :bytes, OpenAI::Internal::Type::ArrayOf[OpenAI::Internal::Type::Unknown] # @!parse # # @return [Array] @@ -80,7 +81,7 @@ class Logprob < OpenAI::BaseModel # # # def initialize(token: nil, bytes: nil, logprob: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/audio/transcription_text_done_event.rb b/lib/openai/models/audio/transcription_text_done_event.rb index b2a78b25..9795995a 100644 --- a/lib/openai/models/audio/transcription_text_done_event.rb +++ b/lib/openai/models/audio/transcription_text_done_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Audio - class TranscriptionTextDoneEvent < OpenAI::BaseModel + class TranscriptionTextDoneEvent < OpenAI::Internal::Type::BaseModel # @!attribute text # The text that was transcribed. # @@ -23,7 +23,8 @@ class TranscriptionTextDoneEvent < OpenAI::BaseModel # with the `include[]` parameter set to `logprobs`. # # @return [Array, nil] - optional :logprobs, -> { OpenAI::ArrayOf[OpenAI::Models::Audio::TranscriptionTextDoneEvent::Logprob] } + optional :logprobs, + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Audio::TranscriptionTextDoneEvent::Logprob] } # @!parse # # @return [Array] @@ -41,9 +42,9 @@ class TranscriptionTextDoneEvent < OpenAI::BaseModel # # # def initialize(text:, logprobs: nil, type: :"transcript.text.done", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class Logprob < OpenAI::BaseModel + class Logprob < OpenAI::Internal::Type::BaseModel # @!attribute [r] token # The token that was used to generate the log probability. # @@ -58,7 +59,7 @@ class Logprob < OpenAI::BaseModel # The bytes that were used to generate the log probability. # # @return [Array, nil] - optional :bytes, OpenAI::ArrayOf[OpenAI::Unknown] + optional :bytes, OpenAI::Internal::Type::ArrayOf[OpenAI::Internal::Type::Unknown] # @!parse # # @return [Array] @@ -81,7 +82,7 @@ class Logprob < OpenAI::BaseModel # # # def initialize(token: nil, bytes: nil, logprob: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/audio/transcription_verbose.rb b/lib/openai/models/audio/transcription_verbose.rb index 3b66f702..1dfb5931 100644 --- a/lib/openai/models/audio/transcription_verbose.rb +++ b/lib/openai/models/audio/transcription_verbose.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Audio - class TranscriptionVerbose < OpenAI::BaseModel + class TranscriptionVerbose < OpenAI::Internal::Type::BaseModel # @!attribute duration # The duration of the input audio. # @@ -26,7 +26,7 @@ class TranscriptionVerbose < OpenAI::BaseModel # Segments of the transcribed text and their corresponding details. # # @return [Array, nil] - optional :segments, -> { OpenAI::ArrayOf[OpenAI::Models::Audio::TranscriptionSegment] } + optional :segments, -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Audio::TranscriptionSegment] } # @!parse # # @return [Array] @@ -36,7 +36,7 @@ class TranscriptionVerbose < OpenAI::BaseModel # Extracted words and their corresponding timestamps. # # @return [Array, nil] - optional :words, -> { OpenAI::ArrayOf[OpenAI::Models::Audio::TranscriptionWord] } + optional :words, -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Audio::TranscriptionWord] } # @!parse # # @return [Array] @@ -54,7 +54,7 @@ class TranscriptionVerbose < OpenAI::BaseModel # # # def initialize(duration:, language:, text:, segments: nil, words: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/audio/transcription_word.rb b/lib/openai/models/audio/transcription_word.rb index f331b86c..5df30687 100644 --- a/lib/openai/models/audio/transcription_word.rb +++ b/lib/openai/models/audio/transcription_word.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Audio - class TranscriptionWord < OpenAI::BaseModel + class TranscriptionWord < OpenAI::Internal::Type::BaseModel # @!attribute end_ # End time of the word in seconds. # @@ -29,7 +29,7 @@ class TranscriptionWord < OpenAI::BaseModel # # # def initialize(end_:, start:, word:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/audio/translation.rb b/lib/openai/models/audio/translation.rb index 7b9597ca..c9631757 100644 --- a/lib/openai/models/audio/translation.rb +++ b/lib/openai/models/audio/translation.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Audio - class Translation < OpenAI::BaseModel + class Translation < OpenAI::Internal::Type::BaseModel # @!attribute text # # @return [String] @@ -14,7 +14,7 @@ class Translation < OpenAI::BaseModel # # # def initialize(text:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/audio/translation_create_params.rb b/lib/openai/models/audio/translation_create_params.rb index 57b2727c..93833eb2 100644 --- a/lib/openai/models/audio/translation_create_params.rb +++ b/lib/openai/models/audio/translation_create_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Audio # @see OpenAI::Resources::Audio::Translations#create - class TranslationCreateParams < OpenAI::BaseModel + class TranslationCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -71,12 +71,12 @@ class TranslationCreateParams < OpenAI::BaseModel # # # def initialize(file:, model:, prompt: nil, response_format: nil, temperature: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # ID of the model to use. Only `whisper-1` (which is powered by our open source # Whisper V2 model) is currently available. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String @@ -91,7 +91,7 @@ module Model # The format of the output, in one of these options: `json`, `text`, `srt`, # `verbose_json`, or `vtt`. module ResponseFormat - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum JSON = :json TEXT = :text diff --git a/lib/openai/models/audio/translation_create_response.rb b/lib/openai/models/audio/translation_create_response.rb index 3beab227..f50f6c86 100644 --- a/lib/openai/models/audio/translation_create_response.rb +++ b/lib/openai/models/audio/translation_create_response.rb @@ -5,7 +5,7 @@ module Models module Audio # @see OpenAI::Resources::Audio::Translations#create module TranslationCreateResponse - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant -> { OpenAI::Models::Audio::Translation } diff --git a/lib/openai/models/audio/translation_verbose.rb b/lib/openai/models/audio/translation_verbose.rb index eb6ecfdd..b8a970d0 100644 --- a/lib/openai/models/audio/translation_verbose.rb +++ b/lib/openai/models/audio/translation_verbose.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Audio - class TranslationVerbose < OpenAI::BaseModel + class TranslationVerbose < OpenAI::Internal::Type::BaseModel # @!attribute duration # The duration of the input audio. # @@ -26,7 +26,7 @@ class TranslationVerbose < OpenAI::BaseModel # Segments of the translated text and their corresponding details. # # @return [Array, nil] - optional :segments, -> { OpenAI::ArrayOf[OpenAI::Models::Audio::TranscriptionSegment] } + optional :segments, -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Audio::TranscriptionSegment] } # @!parse # # @return [Array] @@ -40,7 +40,7 @@ class TranslationVerbose < OpenAI::BaseModel # # # def initialize(duration:, language:, text:, segments: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/audio_model.rb b/lib/openai/models/audio_model.rb index d0ffd0c5..4373ee6b 100644 --- a/lib/openai/models/audio_model.rb +++ b/lib/openai/models/audio_model.rb @@ -3,7 +3,7 @@ module OpenAI module Models module AudioModel - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum WHISPER_1 = :"whisper-1" GPT_4O_TRANSCRIBE = :"gpt-4o-transcribe" diff --git a/lib/openai/models/audio_response_format.rb b/lib/openai/models/audio_response_format.rb index 0b6e75b3..0904c657 100644 --- a/lib/openai/models/audio_response_format.rb +++ b/lib/openai/models/audio_response_format.rb @@ -6,7 +6,7 @@ module Models # `verbose_json`, or `vtt`. For `gpt-4o-transcribe` and `gpt-4o-mini-transcribe`, # the only supported format is `json`. module AudioResponseFormat - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum JSON = :json TEXT = :text diff --git a/lib/openai/models/auto_file_chunking_strategy_param.rb b/lib/openai/models/auto_file_chunking_strategy_param.rb index 124d43f7..15d03f11 100644 --- a/lib/openai/models/auto_file_chunking_strategy_param.rb +++ b/lib/openai/models/auto_file_chunking_strategy_param.rb @@ -2,7 +2,7 @@ module OpenAI module Models - class AutoFileChunkingStrategyParam < OpenAI::BaseModel + class AutoFileChunkingStrategyParam < OpenAI::Internal::Type::BaseModel # @!attribute type # Always `auto`. # @@ -17,7 +17,7 @@ class AutoFileChunkingStrategyParam < OpenAI::BaseModel # # # def initialize(type: :auto, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/batch.rb b/lib/openai/models/batch.rb index 5d59baff..be9ede39 100644 --- a/lib/openai/models/batch.rb +++ b/lib/openai/models/batch.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Batches#create - class Batch < OpenAI::BaseModel + class Batch < OpenAI::Internal::Type::BaseModel # @!attribute id # # @return [String] @@ -153,7 +153,7 @@ class Batch < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!attribute [r] output_file_id # The ID of the file containing the outputs of successfully executed requests. @@ -223,13 +223,13 @@ class Batch < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The current status of the batch. # # @see OpenAI::Models::Batch#status module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum VALIDATING = :validating FAILED = :failed @@ -248,11 +248,11 @@ module Status end # @see OpenAI::Models::Batch#errors - class Errors < OpenAI::BaseModel + class Errors < OpenAI::Internal::Type::BaseModel # @!attribute [r] data # # @return [Array, nil] - optional :data, -> { OpenAI::ArrayOf[OpenAI::Models::BatchError] } + optional :data, -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::BatchError] } # @!parse # # @return [Array] @@ -274,7 +274,7 @@ class Errors < OpenAI::BaseModel # # # def initialize(data: nil, object: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/batch_cancel_params.rb b/lib/openai/models/batch_cancel_params.rb index a9fb8c08..4e1871e0 100644 --- a/lib/openai/models/batch_cancel_params.rb +++ b/lib/openai/models/batch_cancel_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Batches#cancel - class BatchCancelParams < OpenAI::BaseModel + class BatchCancelParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -13,7 +13,7 @@ class BatchCancelParams < OpenAI::BaseModel # # # def initialize(request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/batch_create_params.rb b/lib/openai/models/batch_create_params.rb index 04139051..1549f66e 100644 --- a/lib/openai/models/batch_create_params.rb +++ b/lib/openai/models/batch_create_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Batches#create - class BatchCreateParams < OpenAI::BaseModel + class BatchCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -47,7 +47,7 @@ class BatchCreateParams < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!parse # # @param completion_window [Symbol, OpenAI::Models::BatchCreateParams::CompletionWindow] @@ -58,12 +58,12 @@ class BatchCreateParams < OpenAI::BaseModel # # # def initialize(completion_window:, endpoint:, input_file_id:, metadata: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The time frame within which the batch should be processed. Currently only `24h` # is supported. module CompletionWindow - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum COMPLETION_WINDOW_24H = :"24h" @@ -79,7 +79,7 @@ module CompletionWindow # are supported. Note that `/v1/embeddings` batches are also restricted to a # maximum of 50,000 embedding inputs across all requests in the batch. module Endpoint - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum V1_RESPONSES = :"/v1/responses" V1_CHAT_COMPLETIONS = :"/v1/chat/completions" diff --git a/lib/openai/models/batch_error.rb b/lib/openai/models/batch_error.rb index 5333a1ab..3d31ed4a 100644 --- a/lib/openai/models/batch_error.rb +++ b/lib/openai/models/batch_error.rb @@ -2,7 +2,7 @@ module OpenAI module Models - class BatchError < OpenAI::BaseModel + class BatchError < OpenAI::Internal::Type::BaseModel # @!attribute [r] code # An error code identifying the error type. # @@ -43,7 +43,7 @@ class BatchError < OpenAI::BaseModel # # # def initialize(code: nil, line: nil, message: nil, param: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/batch_list_params.rb b/lib/openai/models/batch_list_params.rb index 0dc270ae..651111c7 100644 --- a/lib/openai/models/batch_list_params.rb +++ b/lib/openai/models/batch_list_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Batches#list - class BatchListParams < OpenAI::BaseModel + class BatchListParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -39,7 +39,7 @@ class BatchListParams < OpenAI::BaseModel # # # def initialize(after: nil, limit: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/batch_request_counts.rb b/lib/openai/models/batch_request_counts.rb index 2029a404..92ab5316 100644 --- a/lib/openai/models/batch_request_counts.rb +++ b/lib/openai/models/batch_request_counts.rb @@ -2,7 +2,7 @@ module OpenAI module Models - class BatchRequestCounts < OpenAI::BaseModel + class BatchRequestCounts < OpenAI::Internal::Type::BaseModel # @!attribute completed # Number of requests that have been completed successfully. # @@ -30,7 +30,7 @@ class BatchRequestCounts < OpenAI::BaseModel # # # def initialize(completed:, failed:, total:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/batch_retrieve_params.rb b/lib/openai/models/batch_retrieve_params.rb index 695466e8..fac34345 100644 --- a/lib/openai/models/batch_retrieve_params.rb +++ b/lib/openai/models/batch_retrieve_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Batches#retrieve - class BatchRetrieveParams < OpenAI::BaseModel + class BatchRetrieveParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -13,7 +13,7 @@ class BatchRetrieveParams < OpenAI::BaseModel # # # def initialize(request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/assistant.rb b/lib/openai/models/beta/assistant.rb index 9adc23a7..727f24f2 100644 --- a/lib/openai/models/beta/assistant.rb +++ b/lib/openai/models/beta/assistant.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta # @see OpenAI::Resources::Beta::Assistants#create - class Assistant < OpenAI::BaseModel + class Assistant < OpenAI::Internal::Type::BaseModel # @!attribute id # The identifier, which can be referenced in API endpoints. # @@ -39,7 +39,7 @@ class Assistant < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - required :metadata, OpenAI::HashOf[String], nil?: true + required :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!attribute model # ID of the model to use. You can use the @@ -69,7 +69,7 @@ class Assistant < OpenAI::BaseModel # `function`. # # @return [Array] - required :tools, -> { OpenAI::ArrayOf[union: OpenAI::Models::Beta::AssistantTool] } + required :tools, -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::AssistantTool] } # @!attribute response_format # Specifies the format that the model must output. Compatible with @@ -159,10 +159,10 @@ class Assistant < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::Assistant#tool_resources - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel # @!attribute [r] code_interpreter # # @return [OpenAI::Models::Beta::Assistant::ToolResources::CodeInterpreter, nil] @@ -192,17 +192,17 @@ class ToolResources < OpenAI::BaseModel # # # def initialize(code_interpreter: nil, file_search: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::Assistant::ToolResources#code_interpreter - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel # @!attribute [r] file_ids # A list of [file](https://platform.openai.com/docs/api-reference/files) IDs made # available to the `code_interpreter`` tool. There can be a maximum of 20 files # associated with the tool. # # @return [Array, nil] - optional :file_ids, OpenAI::ArrayOf[String] + optional :file_ids, OpenAI::Internal::Type::ArrayOf[String] # @!parse # # @return [Array] @@ -213,11 +213,11 @@ class CodeInterpreter < OpenAI::BaseModel # # # def initialize(file_ids: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @see OpenAI::Models::Beta::Assistant::ToolResources#file_search - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # @!attribute [r] vector_store_ids # The ID of the # [vector store](https://platform.openai.com/docs/api-reference/vector-stores/object) @@ -225,7 +225,7 @@ class FileSearch < OpenAI::BaseModel # the assistant. # # @return [Array, nil] - optional :vector_store_ids, OpenAI::ArrayOf[String] + optional :vector_store_ids, OpenAI::Internal::Type::ArrayOf[String] # @!parse # # @return [Array] @@ -236,7 +236,7 @@ class FileSearch < OpenAI::BaseModel # # # def initialize(vector_store_ids: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/assistant_create_params.rb b/lib/openai/models/beta/assistant_create_params.rb index d6443c05..2b426970 100644 --- a/lib/openai/models/beta/assistant_create_params.rb +++ b/lib/openai/models/beta/assistant_create_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta # @see OpenAI::Resources::Beta::Assistants#create - class AssistantCreateParams < OpenAI::BaseModel + class AssistantCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -41,7 +41,7 @@ class AssistantCreateParams < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!attribute name # The name of the assistant. The maximum length is 256 characters. @@ -108,7 +108,7 @@ class AssistantCreateParams < OpenAI::BaseModel # `function`. # # @return [Array, nil] - optional :tools, -> { OpenAI::ArrayOf[union: OpenAI::Models::Beta::AssistantTool] } + optional :tools, -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::AssistantTool] } # @!parse # # @return [Array] @@ -156,7 +156,7 @@ class AssistantCreateParams < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # ID of the model to use. You can use the # [List models](https://platform.openai.com/docs/api-reference/models/list) API to @@ -164,7 +164,7 @@ class AssistantCreateParams < OpenAI::BaseModel # [Model overview](https://platform.openai.com/docs/models) for descriptions of # them. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String @@ -176,7 +176,7 @@ module Model # def self.variants; end end - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel # @!attribute [r] code_interpreter # # @return [OpenAI::Models::Beta::AssistantCreateParams::ToolResources::CodeInterpreter, nil] @@ -207,17 +207,17 @@ class ToolResources < OpenAI::BaseModel # # # def initialize(code_interpreter: nil, file_search: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::AssistantCreateParams::ToolResources#code_interpreter - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel # @!attribute [r] file_ids # A list of [file](https://platform.openai.com/docs/api-reference/files) IDs made # available to the `code_interpreter` tool. There can be a maximum of 20 files # associated with the tool. # # @return [Array, nil] - optional :file_ids, OpenAI::ArrayOf[String] + optional :file_ids, OpenAI::Internal::Type::ArrayOf[String] # @!parse # # @return [Array] @@ -228,11 +228,11 @@ class CodeInterpreter < OpenAI::BaseModel # # # def initialize(file_ids: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @see OpenAI::Models::Beta::AssistantCreateParams::ToolResources#file_search - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # @!attribute [r] vector_store_ids # The # [vector store](https://platform.openai.com/docs/api-reference/vector-stores/object) @@ -240,7 +240,7 @@ class FileSearch < OpenAI::BaseModel # the assistant. # # @return [Array, nil] - optional :vector_store_ids, OpenAI::ArrayOf[String] + optional :vector_store_ids, OpenAI::Internal::Type::ArrayOf[String] # @!parse # # @return [Array] @@ -254,7 +254,7 @@ class FileSearch < OpenAI::BaseModel # # @return [Array, nil] optional :vector_stores, - -> { OpenAI::ArrayOf[OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch::VectorStore] } + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch::VectorStore] } # @!parse # # @return [Array] @@ -266,9 +266,9 @@ class FileSearch < OpenAI::BaseModel # # # def initialize(vector_store_ids: nil, vector_stores: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class VectorStore < OpenAI::BaseModel + class VectorStore < OpenAI::Internal::Type::BaseModel # @!attribute [r] chunking_strategy # The chunking strategy used to chunk the file(s). If not set, will use the `auto` # strategy. @@ -287,7 +287,7 @@ class VectorStore < OpenAI::BaseModel # store. # # @return [Array, nil] - optional :file_ids, OpenAI::ArrayOf[String] + optional :file_ids, OpenAI::Internal::Type::ArrayOf[String] # @!parse # # @return [Array] @@ -302,7 +302,7 @@ class VectorStore < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!parse # # @param chunking_strategy [OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Auto, OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static] @@ -311,14 +311,14 @@ class VectorStore < OpenAI::BaseModel # # # def initialize(chunking_strategy: nil, file_ids: nil, metadata: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The chunking strategy used to chunk the file(s). If not set, will use the `auto` # strategy. # # @see OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch::VectorStore#chunking_strategy module ChunkingStrategy - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type @@ -329,7 +329,7 @@ module ChunkingStrategy variant :static, -> { OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static } - class Auto < OpenAI::BaseModel + class Auto < OpenAI::Internal::Type::BaseModel # @!attribute type # Always `auto`. # @@ -344,10 +344,10 @@ class Auto < OpenAI::BaseModel # # # def initialize(type: :auto, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class Static < OpenAI::BaseModel + class Static < OpenAI::Internal::Type::BaseModel # @!attribute static # # @return [OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static::Static] @@ -366,10 +366,10 @@ class Static < OpenAI::BaseModel # # # def initialize(static:, type: :static, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static#static - class Static < OpenAI::BaseModel + class Static < OpenAI::Internal::Type::BaseModel # @!attribute chunk_overlap_tokens # The number of tokens that overlap between chunks. The default value is `400`. # @@ -391,7 +391,7 @@ class Static < OpenAI::BaseModel # # # def initialize(chunk_overlap_tokens:, max_chunk_size_tokens:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end diff --git a/lib/openai/models/beta/assistant_delete_params.rb b/lib/openai/models/beta/assistant_delete_params.rb index 1414cb2b..aed0abc7 100644 --- a/lib/openai/models/beta/assistant_delete_params.rb +++ b/lib/openai/models/beta/assistant_delete_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta # @see OpenAI::Resources::Beta::Assistants#delete - class AssistantDeleteParams < OpenAI::BaseModel + class AssistantDeleteParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -14,7 +14,7 @@ class AssistantDeleteParams < OpenAI::BaseModel # # # def initialize(request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/assistant_deleted.rb b/lib/openai/models/beta/assistant_deleted.rb index 4793b0ae..f4af5c72 100644 --- a/lib/openai/models/beta/assistant_deleted.rb +++ b/lib/openai/models/beta/assistant_deleted.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta # @see OpenAI::Resources::Beta::Assistants#delete - class AssistantDeleted < OpenAI::BaseModel + class AssistantDeleted < OpenAI::Internal::Type::BaseModel # @!attribute id # # @return [String] @@ -13,7 +13,7 @@ class AssistantDeleted < OpenAI::BaseModel # @!attribute deleted # # @return [Boolean] - required :deleted, OpenAI::BooleanModel + required :deleted, OpenAI::Internal::Type::BooleanModel # @!attribute object # @@ -27,7 +27,7 @@ class AssistantDeleted < OpenAI::BaseModel # # # def initialize(id:, deleted:, object: :"assistant.deleted", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/assistant_list_params.rb b/lib/openai/models/beta/assistant_list_params.rb index 1385f253..c01b9f64 100644 --- a/lib/openai/models/beta/assistant_list_params.rb +++ b/lib/openai/models/beta/assistant_list_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta # @see OpenAI::Resources::Beta::Assistants#list - class AssistantListParams < OpenAI::BaseModel + class AssistantListParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -66,12 +66,12 @@ class AssistantListParams < OpenAI::BaseModel # # # def initialize(after: nil, before: nil, limit: nil, order: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Sort order by the `created_at` timestamp of the objects. `asc` for ascending # order and `desc` for descending order. module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASC = :asc DESC = :desc diff --git a/lib/openai/models/beta/assistant_response_format_option.rb b/lib/openai/models/beta/assistant_response_format_option.rb index 48b15aaa..2c797ac5 100644 --- a/lib/openai/models/beta/assistant_response_format_option.rb +++ b/lib/openai/models/beta/assistant_response_format_option.rb @@ -24,7 +24,7 @@ module Beta # indicates the generation exceeded `max_tokens` or the conversation exceeded the # max context length. module AssistantResponseFormatOption - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # `auto` is the default value variant const: :auto diff --git a/lib/openai/models/beta/assistant_retrieve_params.rb b/lib/openai/models/beta/assistant_retrieve_params.rb index 1c58ceea..1a86f690 100644 --- a/lib/openai/models/beta/assistant_retrieve_params.rb +++ b/lib/openai/models/beta/assistant_retrieve_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta # @see OpenAI::Resources::Beta::Assistants#retrieve - class AssistantRetrieveParams < OpenAI::BaseModel + class AssistantRetrieveParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -14,7 +14,7 @@ class AssistantRetrieveParams < OpenAI::BaseModel # # # def initialize(request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/assistant_stream_event.rb b/lib/openai/models/beta/assistant_stream_event.rb index 734ad6c4..9bd4aec3 100644 --- a/lib/openai/models/beta/assistant_stream_event.rb +++ b/lib/openai/models/beta/assistant_stream_event.rb @@ -24,7 +24,7 @@ module Beta # [Assistants API quickstart](https://platform.openai.com/docs/assistants/overview) # to learn how to integrate the Assistants API with streaming. module AssistantStreamEvent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :event @@ -109,7 +109,7 @@ module AssistantStreamEvent # Occurs when an [error](https://platform.openai.com/docs/guides/error-codes#api-errors) occurs. This can happen due to an internal server error or a timeout. variant :error, -> { OpenAI::Models::Beta::AssistantStreamEvent::ErrorEvent } - class ThreadCreated < OpenAI::BaseModel + class ThreadCreated < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a thread that contains # [messages](https://platform.openai.com/docs/api-reference/messages). @@ -126,7 +126,7 @@ class ThreadCreated < OpenAI::BaseModel # Whether to enable input audio transcription. # # @return [Boolean, nil] - optional :enabled, OpenAI::BooleanModel + optional :enabled, OpenAI::Internal::Type::BooleanModel # @!parse # # @return [Boolean] @@ -143,10 +143,10 @@ class ThreadCreated < OpenAI::BaseModel # # # def initialize(data:, enabled: nil, event: :"thread.created", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunCreated < OpenAI::BaseModel + class ThreadRunCreated < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -168,10 +168,10 @@ class ThreadRunCreated < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.created", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunQueued < OpenAI::BaseModel + class ThreadRunQueued < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -193,10 +193,10 @@ class ThreadRunQueued < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.queued", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunInProgress < OpenAI::BaseModel + class ThreadRunInProgress < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -218,10 +218,10 @@ class ThreadRunInProgress < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.in_progress", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunRequiresAction < OpenAI::BaseModel + class ThreadRunRequiresAction < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -243,10 +243,10 @@ class ThreadRunRequiresAction < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.requires_action", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunCompleted < OpenAI::BaseModel + class ThreadRunCompleted < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -268,10 +268,10 @@ class ThreadRunCompleted < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.completed", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunIncomplete < OpenAI::BaseModel + class ThreadRunIncomplete < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -293,10 +293,10 @@ class ThreadRunIncomplete < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.incomplete", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunFailed < OpenAI::BaseModel + class ThreadRunFailed < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -318,10 +318,10 @@ class ThreadRunFailed < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.failed", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunCancelling < OpenAI::BaseModel + class ThreadRunCancelling < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -343,10 +343,10 @@ class ThreadRunCancelling < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.cancelling", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunCancelled < OpenAI::BaseModel + class ThreadRunCancelled < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -368,10 +368,10 @@ class ThreadRunCancelled < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.cancelled", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunExpired < OpenAI::BaseModel + class ThreadRunExpired < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -393,10 +393,10 @@ class ThreadRunExpired < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.expired", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunStepCreated < OpenAI::BaseModel + class ThreadRunStepCreated < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a step in execution of a run. # @@ -418,10 +418,10 @@ class ThreadRunStepCreated < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.step.created", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunStepInProgress < OpenAI::BaseModel + class ThreadRunStepInProgress < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a step in execution of a run. # @@ -443,10 +443,10 @@ class ThreadRunStepInProgress < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.step.in_progress", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunStepDelta < OpenAI::BaseModel + class ThreadRunStepDelta < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a run step delta i.e. any changed fields on a run step during # streaming. @@ -469,10 +469,10 @@ class ThreadRunStepDelta < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.step.delta", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunStepCompleted < OpenAI::BaseModel + class ThreadRunStepCompleted < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a step in execution of a run. # @@ -494,10 +494,10 @@ class ThreadRunStepCompleted < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.step.completed", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunStepFailed < OpenAI::BaseModel + class ThreadRunStepFailed < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a step in execution of a run. # @@ -519,10 +519,10 @@ class ThreadRunStepFailed < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.step.failed", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunStepCancelled < OpenAI::BaseModel + class ThreadRunStepCancelled < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a step in execution of a run. # @@ -544,10 +544,10 @@ class ThreadRunStepCancelled < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.step.cancelled", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunStepExpired < OpenAI::BaseModel + class ThreadRunStepExpired < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a step in execution of a run. # @@ -569,10 +569,10 @@ class ThreadRunStepExpired < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.step.expired", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadMessageCreated < OpenAI::BaseModel + class ThreadMessageCreated < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a message within a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -595,10 +595,10 @@ class ThreadMessageCreated < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.message.created", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadMessageInProgress < OpenAI::BaseModel + class ThreadMessageInProgress < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a message within a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -621,10 +621,10 @@ class ThreadMessageInProgress < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.message.in_progress", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadMessageDelta < OpenAI::BaseModel + class ThreadMessageDelta < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a message delta i.e. any changed fields on a message during # streaming. @@ -647,10 +647,10 @@ class ThreadMessageDelta < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.message.delta", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadMessageCompleted < OpenAI::BaseModel + class ThreadMessageCompleted < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a message within a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -673,10 +673,10 @@ class ThreadMessageCompleted < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.message.completed", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadMessageIncomplete < OpenAI::BaseModel + class ThreadMessageIncomplete < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a message within a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -699,10 +699,10 @@ class ThreadMessageIncomplete < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.message.incomplete", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ErrorEvent < OpenAI::BaseModel + class ErrorEvent < OpenAI::Internal::Type::BaseModel # @!attribute data # # @return [OpenAI::Models::ErrorObject] @@ -723,7 +723,7 @@ class ErrorEvent < OpenAI::BaseModel # # # def initialize(data:, event: :error, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @!parse diff --git a/lib/openai/models/beta/assistant_tool.rb b/lib/openai/models/beta/assistant_tool.rb index 2a2c655d..0921b4bb 100644 --- a/lib/openai/models/beta/assistant_tool.rb +++ b/lib/openai/models/beta/assistant_tool.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module AssistantTool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type diff --git a/lib/openai/models/beta/assistant_tool_choice.rb b/lib/openai/models/beta/assistant_tool_choice.rb index 1caf13f7..0c1c8cb9 100644 --- a/lib/openai/models/beta/assistant_tool_choice.rb +++ b/lib/openai/models/beta/assistant_tool_choice.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Beta - class AssistantToolChoice < OpenAI::BaseModel + class AssistantToolChoice < OpenAI::Internal::Type::BaseModel # @!attribute type # The type of the tool. If type is `function`, the function name must be set # @@ -28,13 +28,13 @@ class AssistantToolChoice < OpenAI::BaseModel # # # def initialize(type:, function: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The type of the tool. If type is `function`, the function name must be set # # @see OpenAI::Models::Beta::AssistantToolChoice#type module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum FUNCTION = :function CODE_INTERPRETER = :code_interpreter diff --git a/lib/openai/models/beta/assistant_tool_choice_function.rb b/lib/openai/models/beta/assistant_tool_choice_function.rb index 4e127718..19ca8d48 100644 --- a/lib/openai/models/beta/assistant_tool_choice_function.rb +++ b/lib/openai/models/beta/assistant_tool_choice_function.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Beta - class AssistantToolChoiceFunction < OpenAI::BaseModel + class AssistantToolChoiceFunction < OpenAI::Internal::Type::BaseModel # @!attribute name # The name of the function to call. # @@ -15,7 +15,7 @@ class AssistantToolChoiceFunction < OpenAI::BaseModel # # # def initialize(name:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/assistant_tool_choice_option.rb b/lib/openai/models/beta/assistant_tool_choice_option.rb index 807dc2eb..5445f1a6 100644 --- a/lib/openai/models/beta/assistant_tool_choice_option.rb +++ b/lib/openai/models/beta/assistant_tool_choice_option.rb @@ -11,7 +11,7 @@ module Beta # `{"type": "function", "function": {"name": "my_function"}}` forces the model to # call that tool. module AssistantToolChoiceOption - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # `none` means the model will not call any tools and instead generates a message. `auto` means the model can pick between generating a message or calling one or more tools. `required` means the model must call one or more tools before responding to the user. variant enum: -> { OpenAI::Models::Beta::AssistantToolChoiceOption::Auto } @@ -24,7 +24,7 @@ module AssistantToolChoiceOption # more tools. `required` means the model must call one or more tools before # responding to the user. module Auto - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum NONE = :none AUTO = :auto diff --git a/lib/openai/models/beta/assistant_update_params.rb b/lib/openai/models/beta/assistant_update_params.rb index 11f30474..2fe97acb 100644 --- a/lib/openai/models/beta/assistant_update_params.rb +++ b/lib/openai/models/beta/assistant_update_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta # @see OpenAI::Resources::Beta::Assistants#update - class AssistantUpdateParams < OpenAI::BaseModel + class AssistantUpdateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -31,7 +31,7 @@ class AssistantUpdateParams < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!attribute [r] model # ID of the model to use. You can use the @@ -112,7 +112,7 @@ class AssistantUpdateParams < OpenAI::BaseModel # `function`. # # @return [Array, nil] - optional :tools, -> { OpenAI::ArrayOf[union: OpenAI::Models::Beta::AssistantTool] } + optional :tools, -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::AssistantTool] } # @!parse # # @return [Array] @@ -160,7 +160,7 @@ class AssistantUpdateParams < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # ID of the model to use. You can use the # [List models](https://platform.openai.com/docs/api-reference/models/list) API to @@ -168,7 +168,7 @@ class AssistantUpdateParams < OpenAI::BaseModel # [Model overview](https://platform.openai.com/docs/models) for descriptions of # them. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String @@ -272,7 +272,7 @@ module Model # @!endgroup end - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel # @!attribute [r] code_interpreter # # @return [OpenAI::Models::Beta::AssistantUpdateParams::ToolResources::CodeInterpreter, nil] @@ -303,10 +303,10 @@ class ToolResources < OpenAI::BaseModel # # # def initialize(code_interpreter: nil, file_search: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::AssistantUpdateParams::ToolResources#code_interpreter - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel # @!attribute [r] file_ids # Overrides the list of # [file](https://platform.openai.com/docs/api-reference/files) IDs made available @@ -314,7 +314,7 @@ class CodeInterpreter < OpenAI::BaseModel # with the tool. # # @return [Array, nil] - optional :file_ids, OpenAI::ArrayOf[String] + optional :file_ids, OpenAI::Internal::Type::ArrayOf[String] # @!parse # # @return [Array] @@ -325,11 +325,11 @@ class CodeInterpreter < OpenAI::BaseModel # # # def initialize(file_ids: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @see OpenAI::Models::Beta::AssistantUpdateParams::ToolResources#file_search - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # @!attribute [r] vector_store_ids # Overrides the # [vector store](https://platform.openai.com/docs/api-reference/vector-stores/object) @@ -337,7 +337,7 @@ class FileSearch < OpenAI::BaseModel # the assistant. # # @return [Array, nil] - optional :vector_store_ids, OpenAI::ArrayOf[String] + optional :vector_store_ids, OpenAI::Internal::Type::ArrayOf[String] # @!parse # # @return [Array] @@ -348,7 +348,7 @@ class FileSearch < OpenAI::BaseModel # # # def initialize(vector_store_ids: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/code_interpreter_tool.rb b/lib/openai/models/beta/code_interpreter_tool.rb index 0e226273..3f3013ad 100644 --- a/lib/openai/models/beta/code_interpreter_tool.rb +++ b/lib/openai/models/beta/code_interpreter_tool.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Beta - class CodeInterpreterTool < OpenAI::BaseModel + class CodeInterpreterTool < OpenAI::Internal::Type::BaseModel # @!attribute type # The type of tool being defined: `code_interpreter` # @@ -15,7 +15,7 @@ class CodeInterpreterTool < OpenAI::BaseModel # # # def initialize(type: :code_interpreter, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/file_search_tool.rb b/lib/openai/models/beta/file_search_tool.rb index 9a23f3f2..2428a1a0 100644 --- a/lib/openai/models/beta/file_search_tool.rb +++ b/lib/openai/models/beta/file_search_tool.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Beta - class FileSearchTool < OpenAI::BaseModel + class FileSearchTool < OpenAI::Internal::Type::BaseModel # @!attribute type # The type of tool being defined: `file_search` # @@ -26,10 +26,10 @@ class FileSearchTool < OpenAI::BaseModel # # # def initialize(file_search: nil, type: :file_search, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::FileSearchTool#file_search - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # @!attribute [r] max_num_results # The maximum number of results the file search tool should output. The default is # 20 for `gpt-4*` models and 5 for `gpt-3.5-turbo`. This number should be between @@ -70,10 +70,10 @@ class FileSearch < OpenAI::BaseModel # # # def initialize(max_num_results: nil, ranking_options: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::FileSearchTool::FileSearch#ranking_options - class RankingOptions < OpenAI::BaseModel + class RankingOptions < OpenAI::Internal::Type::BaseModel # @!attribute score_threshold # The score threshold for the file search. All values must be a floating point # number between 0 and 1. @@ -105,14 +105,14 @@ class RankingOptions < OpenAI::BaseModel # # # def initialize(score_threshold:, ranker: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The ranker to use for the file search. If not specified will use the `auto` # ranker. # # @see OpenAI::Models::Beta::FileSearchTool::FileSearch::RankingOptions#ranker module Ranker - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO = :auto DEFAULT_2024_08_21 = :default_2024_08_21 diff --git a/lib/openai/models/beta/function_tool.rb b/lib/openai/models/beta/function_tool.rb index 5b9cec7a..ed879754 100644 --- a/lib/openai/models/beta/function_tool.rb +++ b/lib/openai/models/beta/function_tool.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Beta - class FunctionTool < OpenAI::BaseModel + class FunctionTool < OpenAI::Internal::Type::BaseModel # @!attribute function # # @return [OpenAI::Models::FunctionDefinition] @@ -21,7 +21,7 @@ class FunctionTool < OpenAI::BaseModel # # # def initialize(function:, type: :function, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/message_stream_event.rb b/lib/openai/models/beta/message_stream_event.rb index 6f0bdf04..0657f158 100644 --- a/lib/openai/models/beta/message_stream_event.rb +++ b/lib/openai/models/beta/message_stream_event.rb @@ -7,7 +7,7 @@ module Beta # [message](https://platform.openai.com/docs/api-reference/messages/object) is # created. module MessageStreamEvent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :event @@ -29,7 +29,7 @@ module MessageStreamEvent variant :"thread.message.incomplete", -> { OpenAI::Models::Beta::MessageStreamEvent::ThreadMessageIncomplete } - class ThreadMessageCreated < OpenAI::BaseModel + class ThreadMessageCreated < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a message within a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -52,10 +52,10 @@ class ThreadMessageCreated < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.message.created", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadMessageInProgress < OpenAI::BaseModel + class ThreadMessageInProgress < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a message within a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -78,10 +78,10 @@ class ThreadMessageInProgress < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.message.in_progress", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadMessageDelta < OpenAI::BaseModel + class ThreadMessageDelta < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a message delta i.e. any changed fields on a message during # streaming. @@ -104,10 +104,10 @@ class ThreadMessageDelta < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.message.delta", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadMessageCompleted < OpenAI::BaseModel + class ThreadMessageCompleted < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a message within a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -130,10 +130,10 @@ class ThreadMessageCompleted < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.message.completed", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadMessageIncomplete < OpenAI::BaseModel + class ThreadMessageIncomplete < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a message within a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -156,7 +156,7 @@ class ThreadMessageIncomplete < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.message.incomplete", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @!parse diff --git a/lib/openai/models/beta/run_step_stream_event.rb b/lib/openai/models/beta/run_step_stream_event.rb index a3dc27d6..93f240f7 100644 --- a/lib/openai/models/beta/run_step_stream_event.rb +++ b/lib/openai/models/beta/run_step_stream_event.rb @@ -7,7 +7,7 @@ module Beta # [run step](https://platform.openai.com/docs/api-reference/run-steps/step-object) # is created. module RunStepStreamEvent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :event @@ -35,7 +35,7 @@ module RunStepStreamEvent # Occurs when a [run step](https://platform.openai.com/docs/api-reference/run-steps/step-object) expires. variant :"thread.run.step.expired", -> { OpenAI::Models::Beta::RunStepStreamEvent::ThreadRunStepExpired } - class ThreadRunStepCreated < OpenAI::BaseModel + class ThreadRunStepCreated < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a step in execution of a run. # @@ -57,10 +57,10 @@ class ThreadRunStepCreated < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.step.created", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunStepInProgress < OpenAI::BaseModel + class ThreadRunStepInProgress < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a step in execution of a run. # @@ -82,10 +82,10 @@ class ThreadRunStepInProgress < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.step.in_progress", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunStepDelta < OpenAI::BaseModel + class ThreadRunStepDelta < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a run step delta i.e. any changed fields on a run step during # streaming. @@ -108,10 +108,10 @@ class ThreadRunStepDelta < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.step.delta", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunStepCompleted < OpenAI::BaseModel + class ThreadRunStepCompleted < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a step in execution of a run. # @@ -133,10 +133,10 @@ class ThreadRunStepCompleted < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.step.completed", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunStepFailed < OpenAI::BaseModel + class ThreadRunStepFailed < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a step in execution of a run. # @@ -158,10 +158,10 @@ class ThreadRunStepFailed < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.step.failed", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunStepCancelled < OpenAI::BaseModel + class ThreadRunStepCancelled < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a step in execution of a run. # @@ -183,10 +183,10 @@ class ThreadRunStepCancelled < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.step.cancelled", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunStepExpired < OpenAI::BaseModel + class ThreadRunStepExpired < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a step in execution of a run. # @@ -208,7 +208,7 @@ class ThreadRunStepExpired < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.step.expired", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @!parse diff --git a/lib/openai/models/beta/run_stream_event.rb b/lib/openai/models/beta/run_stream_event.rb index 058cde20..14e5177f 100644 --- a/lib/openai/models/beta/run_stream_event.rb +++ b/lib/openai/models/beta/run_stream_event.rb @@ -6,7 +6,7 @@ module Beta # Occurs when a new # [run](https://platform.openai.com/docs/api-reference/runs/object) is created. module RunStreamEvent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :event @@ -41,7 +41,7 @@ module RunStreamEvent # Occurs when a [run](https://platform.openai.com/docs/api-reference/runs/object) expires. variant :"thread.run.expired", -> { OpenAI::Models::Beta::RunStreamEvent::ThreadRunExpired } - class ThreadRunCreated < OpenAI::BaseModel + class ThreadRunCreated < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -63,10 +63,10 @@ class ThreadRunCreated < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.created", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunQueued < OpenAI::BaseModel + class ThreadRunQueued < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -88,10 +88,10 @@ class ThreadRunQueued < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.queued", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunInProgress < OpenAI::BaseModel + class ThreadRunInProgress < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -113,10 +113,10 @@ class ThreadRunInProgress < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.in_progress", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunRequiresAction < OpenAI::BaseModel + class ThreadRunRequiresAction < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -138,10 +138,10 @@ class ThreadRunRequiresAction < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.requires_action", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunCompleted < OpenAI::BaseModel + class ThreadRunCompleted < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -163,10 +163,10 @@ class ThreadRunCompleted < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.completed", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunIncomplete < OpenAI::BaseModel + class ThreadRunIncomplete < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -188,10 +188,10 @@ class ThreadRunIncomplete < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.incomplete", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunFailed < OpenAI::BaseModel + class ThreadRunFailed < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -213,10 +213,10 @@ class ThreadRunFailed < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.failed", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunCancelling < OpenAI::BaseModel + class ThreadRunCancelling < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -238,10 +238,10 @@ class ThreadRunCancelling < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.cancelling", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunCancelled < OpenAI::BaseModel + class ThreadRunCancelled < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -263,10 +263,10 @@ class ThreadRunCancelled < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.cancelled", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class ThreadRunExpired < OpenAI::BaseModel + class ThreadRunExpired < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). @@ -288,7 +288,7 @@ class ThreadRunExpired < OpenAI::BaseModel # # # def initialize(data:, event: :"thread.run.expired", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @!parse diff --git a/lib/openai/models/beta/thread.rb b/lib/openai/models/beta/thread.rb index c651e7c2..ab9384f2 100644 --- a/lib/openai/models/beta/thread.rb +++ b/lib/openai/models/beta/thread.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta # @see OpenAI::Resources::Beta::Threads#create - class Thread < OpenAI::BaseModel + class Thread < OpenAI::Internal::Type::BaseModel # @!attribute id # The identifier, which can be referenced in API endpoints. # @@ -26,7 +26,7 @@ class Thread < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - required :metadata, OpenAI::HashOf[String], nil?: true + required :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!attribute object # The object type, which is always `thread`. @@ -55,10 +55,10 @@ class Thread < OpenAI::BaseModel # # # def initialize(id:, created_at:, metadata:, tool_resources:, object: :thread, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::Thread#tool_resources - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel # @!attribute [r] code_interpreter # # @return [OpenAI::Models::Beta::Thread::ToolResources::CodeInterpreter, nil] @@ -88,17 +88,17 @@ class ToolResources < OpenAI::BaseModel # # # def initialize(code_interpreter: nil, file_search: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::Thread::ToolResources#code_interpreter - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel # @!attribute [r] file_ids # A list of [file](https://platform.openai.com/docs/api-reference/files) IDs made # available to the `code_interpreter` tool. There can be a maximum of 20 files # associated with the tool. # # @return [Array, nil] - optional :file_ids, OpenAI::ArrayOf[String] + optional :file_ids, OpenAI::Internal::Type::ArrayOf[String] # @!parse # # @return [Array] @@ -109,11 +109,11 @@ class CodeInterpreter < OpenAI::BaseModel # # # def initialize(file_ids: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @see OpenAI::Models::Beta::Thread::ToolResources#file_search - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # @!attribute [r] vector_store_ids # The # [vector store](https://platform.openai.com/docs/api-reference/vector-stores/object) @@ -121,7 +121,7 @@ class FileSearch < OpenAI::BaseModel # the thread. # # @return [Array, nil] - optional :vector_store_ids, OpenAI::ArrayOf[String] + optional :vector_store_ids, OpenAI::Internal::Type::ArrayOf[String] # @!parse # # @return [Array] @@ -132,7 +132,7 @@ class FileSearch < OpenAI::BaseModel # # # def initialize(vector_store_ids: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/thread_create_and_run_params.rb b/lib/openai/models/beta/thread_create_and_run_params.rb index 4425ba3c..03e401db 100644 --- a/lib/openai/models/beta/thread_create_and_run_params.rb +++ b/lib/openai/models/beta/thread_create_and_run_params.rb @@ -6,7 +6,7 @@ module Beta # @see OpenAI::Resources::Beta::Threads#create_and_run # # @see OpenAI::Resources::Beta::Threads#stream_raw - class ThreadCreateAndRunParams < OpenAI::BaseModel + class ThreadCreateAndRunParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -55,7 +55,7 @@ class ThreadCreateAndRunParams < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!attribute model # The ID of the [Model](https://platform.openai.com/docs/api-reference/models) to @@ -72,7 +72,7 @@ class ThreadCreateAndRunParams < OpenAI::BaseModel # during tool use. # # @return [Boolean, nil] - optional :parallel_tool_calls, OpenAI::BooleanModel + optional :parallel_tool_calls, OpenAI::Internal::Type::BooleanModel # @!parse # # @return [Boolean] @@ -149,7 +149,7 @@ class ThreadCreateAndRunParams < OpenAI::BaseModel # # @return [Array, nil] optional :tools, - -> { OpenAI::ArrayOf[union: OpenAI::Models::Beta::ThreadCreateAndRunParams::Tool] }, + -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::ThreadCreateAndRunParams::Tool] }, nil?: true # @!attribute top_p @@ -211,14 +211,14 @@ class ThreadCreateAndRunParams < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The ID of the [Model](https://platform.openai.com/docs/api-reference/models) to # be used to execute this run. If a value is provided here, it will override the # model associated with the assistant. If not, the model associated with the # assistant will be used. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String @@ -230,14 +230,14 @@ module Model # def self.variants; end end - class Thread < OpenAI::BaseModel + class Thread < OpenAI::Internal::Type::BaseModel # @!attribute [r] messages # A list of [messages](https://platform.openai.com/docs/api-reference/messages) to # start the thread with. # # @return [Array, nil] optional :messages, - -> { OpenAI::ArrayOf[OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::Message] } + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::Message] } # @!parse # # @return [Array] @@ -252,7 +252,7 @@ class Thread < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!attribute tool_resources # A set of resources that are made available to the assistant's tools in this @@ -275,9 +275,9 @@ class Thread < OpenAI::BaseModel # # # def initialize(messages: nil, metadata: nil, tool_resources: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class Message < OpenAI::BaseModel + class Message < OpenAI::Internal::Type::BaseModel # @!attribute content # The text contents of the message. # @@ -300,7 +300,7 @@ class Message < OpenAI::BaseModel # # @return [Array, nil] optional :attachments, - -> { OpenAI::ArrayOf[OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::Message::Attachment] }, + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::Message::Attachment] }, nil?: true # @!attribute metadata @@ -312,7 +312,7 @@ class Message < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!parse # # @param content [String, Array] @@ -322,13 +322,13 @@ class Message < OpenAI::BaseModel # # # def initialize(content:, role:, attachments: nil, metadata: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The text contents of the message. # # @see OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::Message#content module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # The text contents of the message. variant String @@ -341,7 +341,7 @@ module Content # def self.variants; end MessageContentPartParamArray = - OpenAI::ArrayOf[union: -> { OpenAI::Models::Beta::Threads::MessageContentPartParam }] + OpenAI::Internal::Type::ArrayOf[union: -> { OpenAI::Models::Beta::Threads::MessageContentPartParam }] end # The role of the entity that is creating the message. Allowed values include: @@ -353,7 +353,7 @@ module Content # # @see OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::Message#role module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum USER = :user ASSISTANT = :assistant @@ -365,7 +365,7 @@ module Role # def self.values; end end - class Attachment < OpenAI::BaseModel + class Attachment < OpenAI::Internal::Type::BaseModel # @!attribute [r] file_id # The ID of the file to attach to the message. # @@ -381,7 +381,7 @@ class Attachment < OpenAI::BaseModel # # @return [Array, nil] optional :tools, - -> { OpenAI::ArrayOf[union: OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::Message::Attachment::Tool] } + -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::Message::Attachment::Tool] } # @!parse # # @return [Array] @@ -393,10 +393,10 @@ class Attachment < OpenAI::BaseModel # # # def initialize(file_id: nil, tools: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void module Tool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type @@ -405,7 +405,7 @@ module Tool variant :file_search, -> { OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::Message::Attachment::Tool::FileSearch } - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # @!attribute type # The type of tool being defined: `file_search` # @@ -417,7 +417,7 @@ class FileSearch < OpenAI::BaseModel # # # def initialize(type: :file_search, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @!parse @@ -428,7 +428,7 @@ class FileSearch < OpenAI::BaseModel end # @see OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread#tool_resources - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel # @!attribute [r] code_interpreter # # @return [OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::CodeInterpreter, nil] @@ -460,17 +460,17 @@ class ToolResources < OpenAI::BaseModel # # # def initialize(code_interpreter: nil, file_search: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources#code_interpreter - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel # @!attribute [r] file_ids # A list of [file](https://platform.openai.com/docs/api-reference/files) IDs made # available to the `code_interpreter` tool. There can be a maximum of 20 files # associated with the tool. # # @return [Array, nil] - optional :file_ids, OpenAI::ArrayOf[String] + optional :file_ids, OpenAI::Internal::Type::ArrayOf[String] # @!parse # # @return [Array] @@ -481,11 +481,11 @@ class CodeInterpreter < OpenAI::BaseModel # # # def initialize(file_ids: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @see OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources#file_search - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # @!attribute [r] vector_store_ids # The # [vector store](https://platform.openai.com/docs/api-reference/vector-stores/object) @@ -493,7 +493,7 @@ class FileSearch < OpenAI::BaseModel # the thread. # # @return [Array, nil] - optional :vector_store_ids, OpenAI::ArrayOf[String] + optional :vector_store_ids, OpenAI::Internal::Type::ArrayOf[String] # @!parse # # @return [Array] @@ -507,7 +507,7 @@ class FileSearch < OpenAI::BaseModel # # @return [Array, nil] optional :vector_stores, - -> { OpenAI::ArrayOf[OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch::VectorStore] } + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch::VectorStore] } # @!parse # # @return [Array] @@ -519,9 +519,9 @@ class FileSearch < OpenAI::BaseModel # # # def initialize(vector_store_ids: nil, vector_stores: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class VectorStore < OpenAI::BaseModel + class VectorStore < OpenAI::Internal::Type::BaseModel # @!attribute [r] chunking_strategy # The chunking strategy used to chunk the file(s). If not set, will use the `auto` # strategy. @@ -540,7 +540,7 @@ class VectorStore < OpenAI::BaseModel # store. # # @return [Array, nil] - optional :file_ids, OpenAI::ArrayOf[String] + optional :file_ids, OpenAI::Internal::Type::ArrayOf[String] # @!parse # # @return [Array] @@ -555,7 +555,7 @@ class VectorStore < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!parse # # @param chunking_strategy [OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Auto, OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static] @@ -564,14 +564,14 @@ class VectorStore < OpenAI::BaseModel # # # def initialize(chunking_strategy: nil, file_ids: nil, metadata: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The chunking strategy used to chunk the file(s). If not set, will use the `auto` # strategy. # # @see OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch::VectorStore#chunking_strategy module ChunkingStrategy - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type @@ -582,7 +582,7 @@ module ChunkingStrategy variant :static, -> { OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static } - class Auto < OpenAI::BaseModel + class Auto < OpenAI::Internal::Type::BaseModel # @!attribute type # Always `auto`. # @@ -597,10 +597,10 @@ class Auto < OpenAI::BaseModel # # # def initialize(type: :auto, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class Static < OpenAI::BaseModel + class Static < OpenAI::Internal::Type::BaseModel # @!attribute static # # @return [OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static::Static] @@ -619,10 +619,10 @@ class Static < OpenAI::BaseModel # # # def initialize(static:, type: :static, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static#static - class Static < OpenAI::BaseModel + class Static < OpenAI::Internal::Type::BaseModel # @!attribute chunk_overlap_tokens # The number of tokens that overlap between chunks. The default value is `400`. # @@ -644,7 +644,7 @@ class Static < OpenAI::BaseModel # # # def initialize(chunk_overlap_tokens:, max_chunk_size_tokens:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end @@ -657,7 +657,7 @@ class Static < OpenAI::BaseModel end end - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel # @!attribute [r] code_interpreter # # @return [OpenAI::Models::Beta::ThreadCreateAndRunParams::ToolResources::CodeInterpreter, nil] @@ -688,17 +688,17 @@ class ToolResources < OpenAI::BaseModel # # # def initialize(code_interpreter: nil, file_search: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::ThreadCreateAndRunParams::ToolResources#code_interpreter - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel # @!attribute [r] file_ids # A list of [file](https://platform.openai.com/docs/api-reference/files) IDs made # available to the `code_interpreter` tool. There can be a maximum of 20 files # associated with the tool. # # @return [Array, nil] - optional :file_ids, OpenAI::ArrayOf[String] + optional :file_ids, OpenAI::Internal::Type::ArrayOf[String] # @!parse # # @return [Array] @@ -709,11 +709,11 @@ class CodeInterpreter < OpenAI::BaseModel # # # def initialize(file_ids: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @see OpenAI::Models::Beta::ThreadCreateAndRunParams::ToolResources#file_search - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # @!attribute [r] vector_store_ids # The ID of the # [vector store](https://platform.openai.com/docs/api-reference/vector-stores/object) @@ -721,7 +721,7 @@ class FileSearch < OpenAI::BaseModel # the assistant. # # @return [Array, nil] - optional :vector_store_ids, OpenAI::ArrayOf[String] + optional :vector_store_ids, OpenAI::Internal::Type::ArrayOf[String] # @!parse # # @return [Array] @@ -732,12 +732,12 @@ class FileSearch < OpenAI::BaseModel # # # def initialize(vector_store_ids: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end module Tool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant -> { OpenAI::Models::Beta::CodeInterpreterTool } @@ -750,7 +750,7 @@ module Tool # def self.variants; end end - class TruncationStrategy < OpenAI::BaseModel + class TruncationStrategy < OpenAI::Internal::Type::BaseModel # @!attribute type # The truncation strategy to use for the thread. The default is `auto`. If set to # `last_messages`, the thread will be truncated to the n most recent messages in @@ -776,7 +776,7 @@ class TruncationStrategy < OpenAI::BaseModel # # # def initialize(type:, last_messages: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The truncation strategy to use for the thread. The default is `auto`. If set to # `last_messages`, the thread will be truncated to the n most recent messages in @@ -785,7 +785,7 @@ class TruncationStrategy < OpenAI::BaseModel # # @see OpenAI::Models::Beta::ThreadCreateAndRunParams::TruncationStrategy#type module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO = :auto LAST_MESSAGES = :last_messages diff --git a/lib/openai/models/beta/thread_create_params.rb b/lib/openai/models/beta/thread_create_params.rb index 434c28c5..6e46487b 100644 --- a/lib/openai/models/beta/thread_create_params.rb +++ b/lib/openai/models/beta/thread_create_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta # @see OpenAI::Resources::Beta::Threads#create - class ThreadCreateParams < OpenAI::BaseModel + class ThreadCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -14,7 +14,8 @@ class ThreadCreateParams < OpenAI::BaseModel # start the thread with. # # @return [Array, nil] - optional :messages, -> { OpenAI::ArrayOf[OpenAI::Models::Beta::ThreadCreateParams::Message] } + optional :messages, + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Beta::ThreadCreateParams::Message] } # @!parse # # @return [Array] @@ -29,7 +30,7 @@ class ThreadCreateParams < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!attribute tool_resources # A set of resources that are made available to the assistant's tools in this @@ -48,9 +49,9 @@ class ThreadCreateParams < OpenAI::BaseModel # # # def initialize(messages: nil, metadata: nil, tool_resources: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class Message < OpenAI::BaseModel + class Message < OpenAI::Internal::Type::BaseModel # @!attribute content # The text contents of the message. # @@ -73,7 +74,7 @@ class Message < OpenAI::BaseModel # # @return [Array, nil] optional :attachments, - -> { OpenAI::ArrayOf[OpenAI::Models::Beta::ThreadCreateParams::Message::Attachment] }, + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Beta::ThreadCreateParams::Message::Attachment] }, nil?: true # @!attribute metadata @@ -85,7 +86,7 @@ class Message < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!parse # # @param content [String, Array] @@ -95,13 +96,13 @@ class Message < OpenAI::BaseModel # # # def initialize(content:, role:, attachments: nil, metadata: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The text contents of the message. # # @see OpenAI::Models::Beta::ThreadCreateParams::Message#content module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # The text contents of the message. variant String @@ -114,7 +115,7 @@ module Content # def self.variants; end MessageContentPartParamArray = - OpenAI::ArrayOf[union: -> { OpenAI::Models::Beta::Threads::MessageContentPartParam }] + OpenAI::Internal::Type::ArrayOf[union: -> { OpenAI::Models::Beta::Threads::MessageContentPartParam }] end # The role of the entity that is creating the message. Allowed values include: @@ -126,7 +127,7 @@ module Content # # @see OpenAI::Models::Beta::ThreadCreateParams::Message#role module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum USER = :user ASSISTANT = :assistant @@ -138,7 +139,7 @@ module Role # def self.values; end end - class Attachment < OpenAI::BaseModel + class Attachment < OpenAI::Internal::Type::BaseModel # @!attribute [r] file_id # The ID of the file to attach to the message. # @@ -154,7 +155,7 @@ class Attachment < OpenAI::BaseModel # # @return [Array, nil] optional :tools, - -> { OpenAI::ArrayOf[union: OpenAI::Models::Beta::ThreadCreateParams::Message::Attachment::Tool] } + -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::ThreadCreateParams::Message::Attachment::Tool] } # @!parse # # @return [Array] @@ -166,10 +167,10 @@ class Attachment < OpenAI::BaseModel # # # def initialize(file_id: nil, tools: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void module Tool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type @@ -178,7 +179,7 @@ module Tool variant :file_search, -> { OpenAI::Models::Beta::ThreadCreateParams::Message::Attachment::Tool::FileSearch } - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # @!attribute type # The type of tool being defined: `file_search` # @@ -190,7 +191,7 @@ class FileSearch < OpenAI::BaseModel # # # def initialize(type: :file_search, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @!parse @@ -200,7 +201,7 @@ class FileSearch < OpenAI::BaseModel end end - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel # @!attribute [r] code_interpreter # # @return [OpenAI::Models::Beta::ThreadCreateParams::ToolResources::CodeInterpreter, nil] @@ -231,17 +232,17 @@ class ToolResources < OpenAI::BaseModel # # # def initialize(code_interpreter: nil, file_search: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::ThreadCreateParams::ToolResources#code_interpreter - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel # @!attribute [r] file_ids # A list of [file](https://platform.openai.com/docs/api-reference/files) IDs made # available to the `code_interpreter` tool. There can be a maximum of 20 files # associated with the tool. # # @return [Array, nil] - optional :file_ids, OpenAI::ArrayOf[String] + optional :file_ids, OpenAI::Internal::Type::ArrayOf[String] # @!parse # # @return [Array] @@ -252,11 +253,11 @@ class CodeInterpreter < OpenAI::BaseModel # # # def initialize(file_ids: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @see OpenAI::Models::Beta::ThreadCreateParams::ToolResources#file_search - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # @!attribute [r] vector_store_ids # The # [vector store](https://platform.openai.com/docs/api-reference/vector-stores/object) @@ -264,7 +265,7 @@ class FileSearch < OpenAI::BaseModel # the thread. # # @return [Array, nil] - optional :vector_store_ids, OpenAI::ArrayOf[String] + optional :vector_store_ids, OpenAI::Internal::Type::ArrayOf[String] # @!parse # # @return [Array] @@ -278,7 +279,7 @@ class FileSearch < OpenAI::BaseModel # # @return [Array, nil] optional :vector_stores, - -> { OpenAI::ArrayOf[OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch::VectorStore] } + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch::VectorStore] } # @!parse # # @return [Array] @@ -290,9 +291,9 @@ class FileSearch < OpenAI::BaseModel # # # def initialize(vector_store_ids: nil, vector_stores: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class VectorStore < OpenAI::BaseModel + class VectorStore < OpenAI::Internal::Type::BaseModel # @!attribute [r] chunking_strategy # The chunking strategy used to chunk the file(s). If not set, will use the `auto` # strategy. @@ -311,7 +312,7 @@ class VectorStore < OpenAI::BaseModel # store. # # @return [Array, nil] - optional :file_ids, OpenAI::ArrayOf[String] + optional :file_ids, OpenAI::Internal::Type::ArrayOf[String] # @!parse # # @return [Array] @@ -326,7 +327,7 @@ class VectorStore < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!parse # # @param chunking_strategy [OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Auto, OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static] @@ -335,14 +336,14 @@ class VectorStore < OpenAI::BaseModel # # # def initialize(chunking_strategy: nil, file_ids: nil, metadata: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The chunking strategy used to chunk the file(s). If not set, will use the `auto` # strategy. # # @see OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch::VectorStore#chunking_strategy module ChunkingStrategy - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type @@ -353,7 +354,7 @@ module ChunkingStrategy variant :static, -> { OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static } - class Auto < OpenAI::BaseModel + class Auto < OpenAI::Internal::Type::BaseModel # @!attribute type # Always `auto`. # @@ -368,10 +369,10 @@ class Auto < OpenAI::BaseModel # # # def initialize(type: :auto, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class Static < OpenAI::BaseModel + class Static < OpenAI::Internal::Type::BaseModel # @!attribute static # # @return [OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static::Static] @@ -390,10 +391,10 @@ class Static < OpenAI::BaseModel # # # def initialize(static:, type: :static, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static#static - class Static < OpenAI::BaseModel + class Static < OpenAI::Internal::Type::BaseModel # @!attribute chunk_overlap_tokens # The number of tokens that overlap between chunks. The default value is `400`. # @@ -415,7 +416,7 @@ class Static < OpenAI::BaseModel # # # def initialize(chunk_overlap_tokens:, max_chunk_size_tokens:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end diff --git a/lib/openai/models/beta/thread_delete_params.rb b/lib/openai/models/beta/thread_delete_params.rb index 0febd443..aa6b5341 100644 --- a/lib/openai/models/beta/thread_delete_params.rb +++ b/lib/openai/models/beta/thread_delete_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta # @see OpenAI::Resources::Beta::Threads#delete - class ThreadDeleteParams < OpenAI::BaseModel + class ThreadDeleteParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -14,7 +14,7 @@ class ThreadDeleteParams < OpenAI::BaseModel # # # def initialize(request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/thread_deleted.rb b/lib/openai/models/beta/thread_deleted.rb index 35289d1b..5a290344 100644 --- a/lib/openai/models/beta/thread_deleted.rb +++ b/lib/openai/models/beta/thread_deleted.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta # @see OpenAI::Resources::Beta::Threads#delete - class ThreadDeleted < OpenAI::BaseModel + class ThreadDeleted < OpenAI::Internal::Type::BaseModel # @!attribute id # # @return [String] @@ -13,7 +13,7 @@ class ThreadDeleted < OpenAI::BaseModel # @!attribute deleted # # @return [Boolean] - required :deleted, OpenAI::BooleanModel + required :deleted, OpenAI::Internal::Type::BooleanModel # @!attribute object # @@ -27,7 +27,7 @@ class ThreadDeleted < OpenAI::BaseModel # # # def initialize(id:, deleted:, object: :"thread.deleted", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/thread_retrieve_params.rb b/lib/openai/models/beta/thread_retrieve_params.rb index 7b3c99a3..db2128ba 100644 --- a/lib/openai/models/beta/thread_retrieve_params.rb +++ b/lib/openai/models/beta/thread_retrieve_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta # @see OpenAI::Resources::Beta::Threads#retrieve - class ThreadRetrieveParams < OpenAI::BaseModel + class ThreadRetrieveParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -14,7 +14,7 @@ class ThreadRetrieveParams < OpenAI::BaseModel # # # def initialize(request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/thread_stream_event.rb b/lib/openai/models/beta/thread_stream_event.rb index f24a473b..195b3b77 100644 --- a/lib/openai/models/beta/thread_stream_event.rb +++ b/lib/openai/models/beta/thread_stream_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Beta - class ThreadStreamEvent < OpenAI::BaseModel + class ThreadStreamEvent < OpenAI::Internal::Type::BaseModel # @!attribute data # Represents a thread that contains # [messages](https://platform.openai.com/docs/api-reference/messages). @@ -20,7 +20,7 @@ class ThreadStreamEvent < OpenAI::BaseModel # Whether to enable input audio transcription. # # @return [Boolean, nil] - optional :enabled, OpenAI::BooleanModel + optional :enabled, OpenAI::Internal::Type::BooleanModel # @!parse # # @return [Boolean] @@ -37,7 +37,7 @@ class ThreadStreamEvent < OpenAI::BaseModel # # # def initialize(data:, enabled: nil, event: :"thread.created", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/thread_update_params.rb b/lib/openai/models/beta/thread_update_params.rb index 4ae11357..10ec2820 100644 --- a/lib/openai/models/beta/thread_update_params.rb +++ b/lib/openai/models/beta/thread_update_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta # @see OpenAI::Resources::Beta::Threads#update - class ThreadUpdateParams < OpenAI::BaseModel + class ThreadUpdateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -18,7 +18,7 @@ class ThreadUpdateParams < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!attribute tool_resources # A set of resources that are made available to the assistant's tools in this @@ -36,9 +36,9 @@ class ThreadUpdateParams < OpenAI::BaseModel # # # def initialize(metadata: nil, tool_resources: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel # @!attribute [r] code_interpreter # # @return [OpenAI::Models::Beta::ThreadUpdateParams::ToolResources::CodeInterpreter, nil] @@ -69,17 +69,17 @@ class ToolResources < OpenAI::BaseModel # # # def initialize(code_interpreter: nil, file_search: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::ThreadUpdateParams::ToolResources#code_interpreter - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel # @!attribute [r] file_ids # A list of [file](https://platform.openai.com/docs/api-reference/files) IDs made # available to the `code_interpreter` tool. There can be a maximum of 20 files # associated with the tool. # # @return [Array, nil] - optional :file_ids, OpenAI::ArrayOf[String] + optional :file_ids, OpenAI::Internal::Type::ArrayOf[String] # @!parse # # @return [Array] @@ -90,11 +90,11 @@ class CodeInterpreter < OpenAI::BaseModel # # # def initialize(file_ids: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @see OpenAI::Models::Beta::ThreadUpdateParams::ToolResources#file_search - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # @!attribute [r] vector_store_ids # The # [vector store](https://platform.openai.com/docs/api-reference/vector-stores/object) @@ -102,7 +102,7 @@ class FileSearch < OpenAI::BaseModel # the thread. # # @return [Array, nil] - optional :vector_store_ids, OpenAI::ArrayOf[String] + optional :vector_store_ids, OpenAI::Internal::Type::ArrayOf[String] # @!parse # # @return [Array] @@ -113,7 +113,7 @@ class FileSearch < OpenAI::BaseModel # # # def initialize(vector_store_ids: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/annotation.rb b/lib/openai/models/beta/threads/annotation.rb index b7822e85..0086d3c1 100644 --- a/lib/openai/models/beta/threads/annotation.rb +++ b/lib/openai/models/beta/threads/annotation.rb @@ -8,7 +8,7 @@ module Threads # File associated with the assistant or the message. Generated when the assistant # uses the "file_search" tool to search files. module Annotation - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type diff --git a/lib/openai/models/beta/threads/annotation_delta.rb b/lib/openai/models/beta/threads/annotation_delta.rb index 8bbb2ee1..b5a749fd 100644 --- a/lib/openai/models/beta/threads/annotation_delta.rb +++ b/lib/openai/models/beta/threads/annotation_delta.rb @@ -8,7 +8,7 @@ module Threads # File associated with the assistant or the message. Generated when the assistant # uses the "file_search" tool to search files. module AnnotationDelta - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type diff --git a/lib/openai/models/beta/threads/file_citation_annotation.rb b/lib/openai/models/beta/threads/file_citation_annotation.rb index e471d257..404d9340 100644 --- a/lib/openai/models/beta/threads/file_citation_annotation.rb +++ b/lib/openai/models/beta/threads/file_citation_annotation.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class FileCitationAnnotation < OpenAI::BaseModel + class FileCitationAnnotation < OpenAI::Internal::Type::BaseModel # @!attribute end_index # # @return [Integer] @@ -45,10 +45,10 @@ class FileCitationAnnotation < OpenAI::BaseModel # # # def initialize(end_index:, file_citation:, start_index:, text:, type: :file_citation, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::Threads::FileCitationAnnotation#file_citation - class FileCitation < OpenAI::BaseModel + class FileCitation < OpenAI::Internal::Type::BaseModel # @!attribute file_id # The ID of the specific File the citation is from. # @@ -60,7 +60,7 @@ class FileCitation < OpenAI::BaseModel # # # def initialize(file_id:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/file_citation_delta_annotation.rb b/lib/openai/models/beta/threads/file_citation_delta_annotation.rb index 5d2bf50f..5077a210 100644 --- a/lib/openai/models/beta/threads/file_citation_delta_annotation.rb +++ b/lib/openai/models/beta/threads/file_citation_delta_annotation.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class FileCitationDeltaAnnotation < OpenAI::BaseModel + class FileCitationDeltaAnnotation < OpenAI::Internal::Type::BaseModel # @!attribute index # The index of the annotation in the text content part. # @@ -68,10 +68,10 @@ class FileCitationDeltaAnnotation < OpenAI::BaseModel # # # def initialize(index:, end_index: nil, file_citation: nil, start_index: nil, text: nil, type: :file_citation, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::Threads::FileCitationDeltaAnnotation#file_citation - class FileCitation < OpenAI::BaseModel + class FileCitation < OpenAI::Internal::Type::BaseModel # @!attribute [r] file_id # The ID of the specific File the citation is from. # @@ -98,7 +98,7 @@ class FileCitation < OpenAI::BaseModel # # # def initialize(file_id: nil, quote: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/file_path_annotation.rb b/lib/openai/models/beta/threads/file_path_annotation.rb index fbe60a09..9bf12486 100644 --- a/lib/openai/models/beta/threads/file_path_annotation.rb +++ b/lib/openai/models/beta/threads/file_path_annotation.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class FilePathAnnotation < OpenAI::BaseModel + class FilePathAnnotation < OpenAI::Internal::Type::BaseModel # @!attribute end_index # # @return [Integer] @@ -44,10 +44,10 @@ class FilePathAnnotation < OpenAI::BaseModel # # # def initialize(end_index:, file_path:, start_index:, text:, type: :file_path, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::Threads::FilePathAnnotation#file_path - class FilePath < OpenAI::BaseModel + class FilePath < OpenAI::Internal::Type::BaseModel # @!attribute file_id # The ID of the file that was generated. # @@ -59,7 +59,7 @@ class FilePath < OpenAI::BaseModel # # # def initialize(file_id:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/file_path_delta_annotation.rb b/lib/openai/models/beta/threads/file_path_delta_annotation.rb index 8fe4a8e7..cb1c00a6 100644 --- a/lib/openai/models/beta/threads/file_path_delta_annotation.rb +++ b/lib/openai/models/beta/threads/file_path_delta_annotation.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class FilePathDeltaAnnotation < OpenAI::BaseModel + class FilePathDeltaAnnotation < OpenAI::Internal::Type::BaseModel # @!attribute index # The index of the annotation in the text content part. # @@ -67,10 +67,10 @@ class FilePathDeltaAnnotation < OpenAI::BaseModel # # # def initialize(index:, end_index: nil, file_path: nil, start_index: nil, text: nil, type: :file_path, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::Threads::FilePathDeltaAnnotation#file_path - class FilePath < OpenAI::BaseModel + class FilePath < OpenAI::Internal::Type::BaseModel # @!attribute [r] file_id # The ID of the file that was generated. # @@ -86,7 +86,7 @@ class FilePath < OpenAI::BaseModel # # # def initialize(file_id: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/image_file.rb b/lib/openai/models/beta/threads/image_file.rb index 8f66bf22..bd2e1f25 100644 --- a/lib/openai/models/beta/threads/image_file.rb +++ b/lib/openai/models/beta/threads/image_file.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class ImageFile < OpenAI::BaseModel + class ImageFile < OpenAI::Internal::Type::BaseModel # @!attribute file_id # The [File](https://platform.openai.com/docs/api-reference/files) ID of the image # in the message content. Set `purpose="vision"` when uploading the File if you @@ -30,14 +30,14 @@ class ImageFile < OpenAI::BaseModel # # # def initialize(file_id:, detail: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Specifies the detail level of the image if specified by the user. `low` uses # fewer tokens, you can opt in to high resolution using `high`. # # @see OpenAI::Models::Beta::Threads::ImageFile#detail module Detail - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO = :auto LOW = :low diff --git a/lib/openai/models/beta/threads/image_file_content_block.rb b/lib/openai/models/beta/threads/image_file_content_block.rb index 00121cfe..60fc0073 100644 --- a/lib/openai/models/beta/threads/image_file_content_block.rb +++ b/lib/openai/models/beta/threads/image_file_content_block.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class ImageFileContentBlock < OpenAI::BaseModel + class ImageFileContentBlock < OpenAI::Internal::Type::BaseModel # @!attribute image_file # # @return [OpenAI::Models::Beta::Threads::ImageFile] @@ -25,7 +25,7 @@ class ImageFileContentBlock < OpenAI::BaseModel # # # def initialize(image_file:, type: :image_file, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/image_file_delta.rb b/lib/openai/models/beta/threads/image_file_delta.rb index cb52ad4b..4d654fe4 100644 --- a/lib/openai/models/beta/threads/image_file_delta.rb +++ b/lib/openai/models/beta/threads/image_file_delta.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class ImageFileDelta < OpenAI::BaseModel + class ImageFileDelta < OpenAI::Internal::Type::BaseModel # @!attribute [r] detail # Specifies the detail level of the image if specified by the user. `low` uses # fewer tokens, you can opt in to high resolution using `high`. @@ -34,14 +34,14 @@ class ImageFileDelta < OpenAI::BaseModel # # # def initialize(detail: nil, file_id: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Specifies the detail level of the image if specified by the user. `low` uses # fewer tokens, you can opt in to high resolution using `high`. # # @see OpenAI::Models::Beta::Threads::ImageFileDelta#detail module Detail - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO = :auto LOW = :low diff --git a/lib/openai/models/beta/threads/image_file_delta_block.rb b/lib/openai/models/beta/threads/image_file_delta_block.rb index c33babbb..51118d76 100644 --- a/lib/openai/models/beta/threads/image_file_delta_block.rb +++ b/lib/openai/models/beta/threads/image_file_delta_block.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class ImageFileDeltaBlock < OpenAI::BaseModel + class ImageFileDeltaBlock < OpenAI::Internal::Type::BaseModel # @!attribute index # The index of the content part in the message. # @@ -36,7 +36,7 @@ class ImageFileDeltaBlock < OpenAI::BaseModel # # # def initialize(index:, image_file: nil, type: :image_file, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/image_url.rb b/lib/openai/models/beta/threads/image_url.rb index a38ccdbd..c5db8e23 100644 --- a/lib/openai/models/beta/threads/image_url.rb +++ b/lib/openai/models/beta/threads/image_url.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class ImageURL < OpenAI::BaseModel + class ImageURL < OpenAI::Internal::Type::BaseModel # @!attribute url # The external URL of the image, must be a supported image types: jpeg, jpg, png, # gif, webp. @@ -29,14 +29,14 @@ class ImageURL < OpenAI::BaseModel # # # def initialize(url:, detail: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Specifies the detail level of the image. `low` uses fewer tokens, you can opt in # to high resolution using `high`. Default value is `auto` # # @see OpenAI::Models::Beta::Threads::ImageURL#detail module Detail - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO = :auto LOW = :low diff --git a/lib/openai/models/beta/threads/image_url_content_block.rb b/lib/openai/models/beta/threads/image_url_content_block.rb index b800af73..326f1c1d 100644 --- a/lib/openai/models/beta/threads/image_url_content_block.rb +++ b/lib/openai/models/beta/threads/image_url_content_block.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class ImageURLContentBlock < OpenAI::BaseModel + class ImageURLContentBlock < OpenAI::Internal::Type::BaseModel # @!attribute image_url # # @return [OpenAI::Models::Beta::Threads::ImageURL] @@ -24,7 +24,7 @@ class ImageURLContentBlock < OpenAI::BaseModel # # # def initialize(image_url:, type: :image_url, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/image_url_delta.rb b/lib/openai/models/beta/threads/image_url_delta.rb index 6c80622a..0c69ce94 100644 --- a/lib/openai/models/beta/threads/image_url_delta.rb +++ b/lib/openai/models/beta/threads/image_url_delta.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class ImageURLDelta < OpenAI::BaseModel + class ImageURLDelta < OpenAI::Internal::Type::BaseModel # @!attribute [r] detail # Specifies the detail level of the image. `low` uses fewer tokens, you can opt in # to high resolution using `high`. @@ -33,14 +33,14 @@ class ImageURLDelta < OpenAI::BaseModel # # # def initialize(detail: nil, url: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Specifies the detail level of the image. `low` uses fewer tokens, you can opt in # to high resolution using `high`. # # @see OpenAI::Models::Beta::Threads::ImageURLDelta#detail module Detail - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO = :auto LOW = :low diff --git a/lib/openai/models/beta/threads/image_url_delta_block.rb b/lib/openai/models/beta/threads/image_url_delta_block.rb index 59f49efb..950a87c1 100644 --- a/lib/openai/models/beta/threads/image_url_delta_block.rb +++ b/lib/openai/models/beta/threads/image_url_delta_block.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class ImageURLDeltaBlock < OpenAI::BaseModel + class ImageURLDeltaBlock < OpenAI::Internal::Type::BaseModel # @!attribute index # The index of the content part in the message. # @@ -35,7 +35,7 @@ class ImageURLDeltaBlock < OpenAI::BaseModel # # # def initialize(index:, image_url: nil, type: :image_url, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/message.rb b/lib/openai/models/beta/threads/message.rb index ded2bdd4..4987ff90 100644 --- a/lib/openai/models/beta/threads/message.rb +++ b/lib/openai/models/beta/threads/message.rb @@ -5,7 +5,7 @@ module Models module Beta module Threads # @see OpenAI::Resources::Beta::Threads::Messages#create - class Message < OpenAI::BaseModel + class Message < OpenAI::Internal::Type::BaseModel # @!attribute id # The identifier, which can be referenced in API endpoints. # @@ -25,7 +25,7 @@ class Message < OpenAI::BaseModel # # @return [Array, nil] required :attachments, - -> { OpenAI::ArrayOf[OpenAI::Models::Beta::Threads::Message::Attachment] }, + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Beta::Threads::Message::Attachment] }, nil?: true # @!attribute completed_at @@ -38,7 +38,8 @@ class Message < OpenAI::BaseModel # The content of the message in array of text and/or images. # # @return [Array] - required :content, -> { OpenAI::ArrayOf[union: OpenAI::Models::Beta::Threads::MessageContent] } + required :content, + -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::Threads::MessageContent] } # @!attribute created_at # The Unix timestamp (in seconds) for when the message was created. @@ -67,7 +68,7 @@ class Message < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - required :metadata, OpenAI::HashOf[String], nil?: true + required :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!attribute object # The object type, which is always `thread.message`. @@ -142,9 +143,9 @@ class Message < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class Attachment < OpenAI::BaseModel + class Attachment < OpenAI::Internal::Type::BaseModel # @!attribute [r] file_id # The ID of the file to attach to the message. # @@ -159,7 +160,8 @@ class Attachment < OpenAI::BaseModel # The tools to add this file to. # # @return [Array, nil] - optional :tools, -> { OpenAI::ArrayOf[union: OpenAI::Models::Beta::Threads::Message::Attachment::Tool] } + optional :tools, + -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::Threads::Message::Attachment::Tool] } # @!parse # # @return [Array] @@ -171,16 +173,16 @@ class Attachment < OpenAI::BaseModel # # # def initialize(file_id: nil, tools: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void module Tool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant -> { OpenAI::Models::Beta::CodeInterpreterTool } variant -> { OpenAI::Models::Beta::Threads::Message::Attachment::Tool::AssistantToolsFileSearchTypeOnly } - class AssistantToolsFileSearchTypeOnly < OpenAI::BaseModel + class AssistantToolsFileSearchTypeOnly < OpenAI::Internal::Type::BaseModel # @!attribute type # The type of tool being defined: `file_search` # @@ -192,7 +194,7 @@ class AssistantToolsFileSearchTypeOnly < OpenAI::BaseModel # # # def initialize(type: :file_search, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @!parse @@ -202,7 +204,7 @@ class AssistantToolsFileSearchTypeOnly < OpenAI::BaseModel end # @see OpenAI::Models::Beta::Threads::Message#incomplete_details - class IncompleteDetails < OpenAI::BaseModel + class IncompleteDetails < OpenAI::Internal::Type::BaseModel # @!attribute reason # The reason the message is incomplete. # @@ -216,13 +218,13 @@ class IncompleteDetails < OpenAI::BaseModel # # # def initialize(reason:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The reason the message is incomplete. # # @see OpenAI::Models::Beta::Threads::Message::IncompleteDetails#reason module Reason - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum CONTENT_FILTER = :content_filter MAX_TOKENS = :max_tokens @@ -242,7 +244,7 @@ module Reason # # @see OpenAI::Models::Beta::Threads::Message#role module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum USER = :user ASSISTANT = :assistant @@ -259,7 +261,7 @@ module Role # # @see OpenAI::Models::Beta::Threads::Message#status module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS = :in_progress INCOMPLETE = :incomplete diff --git a/lib/openai/models/beta/threads/message_content.rb b/lib/openai/models/beta/threads/message_content.rb index 4e4280f1..793bbb1d 100644 --- a/lib/openai/models/beta/threads/message_content.rb +++ b/lib/openai/models/beta/threads/message_content.rb @@ -7,7 +7,7 @@ module Threads # References an image [File](https://platform.openai.com/docs/api-reference/files) # in the content of a message. module MessageContent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type diff --git a/lib/openai/models/beta/threads/message_content_delta.rb b/lib/openai/models/beta/threads/message_content_delta.rb index 4cbf92ef..0d6aafd8 100644 --- a/lib/openai/models/beta/threads/message_content_delta.rb +++ b/lib/openai/models/beta/threads/message_content_delta.rb @@ -7,7 +7,7 @@ module Threads # References an image [File](https://platform.openai.com/docs/api-reference/files) # in the content of a message. module MessageContentDelta - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type diff --git a/lib/openai/models/beta/threads/message_content_part_param.rb b/lib/openai/models/beta/threads/message_content_part_param.rb index 20ea5568..33049950 100644 --- a/lib/openai/models/beta/threads/message_content_part_param.rb +++ b/lib/openai/models/beta/threads/message_content_part_param.rb @@ -7,7 +7,7 @@ module Threads # References an image [File](https://platform.openai.com/docs/api-reference/files) # in the content of a message. module MessageContentPartParam - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type diff --git a/lib/openai/models/beta/threads/message_create_params.rb b/lib/openai/models/beta/threads/message_create_params.rb index 5a15f269..dfabf003 100644 --- a/lib/openai/models/beta/threads/message_create_params.rb +++ b/lib/openai/models/beta/threads/message_create_params.rb @@ -5,7 +5,7 @@ module Models module Beta module Threads # @see OpenAI::Resources::Beta::Threads::Messages#create - class MessageCreateParams < OpenAI::BaseModel + class MessageCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -32,7 +32,7 @@ class MessageCreateParams < OpenAI::BaseModel # # @return [Array, nil] optional :attachments, - -> { OpenAI::ArrayOf[OpenAI::Models::Beta::Threads::MessageCreateParams::Attachment] }, + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Beta::Threads::MessageCreateParams::Attachment] }, nil?: true # @!attribute metadata @@ -44,7 +44,7 @@ class MessageCreateParams < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!parse # # @param content [String, Array] @@ -55,11 +55,11 @@ class MessageCreateParams < OpenAI::BaseModel # # # def initialize(content:, role:, attachments: nil, metadata: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The text contents of the message. module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # The text contents of the message. variant String @@ -72,7 +72,7 @@ module Content # def self.variants; end MessageContentPartParamArray = - OpenAI::ArrayOf[union: -> { OpenAI::Models::Beta::Threads::MessageContentPartParam }] + OpenAI::Internal::Type::ArrayOf[union: -> { OpenAI::Models::Beta::Threads::MessageContentPartParam }] end # The role of the entity that is creating the message. Allowed values include: @@ -82,7 +82,7 @@ module Content # - `assistant`: Indicates the message is generated by the assistant. Use this # value to insert messages from the assistant into the conversation. module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum USER = :user ASSISTANT = :assistant @@ -94,7 +94,7 @@ module Role # def self.values; end end - class Attachment < OpenAI::BaseModel + class Attachment < OpenAI::Internal::Type::BaseModel # @!attribute [r] file_id # The ID of the file to attach to the message. # @@ -110,7 +110,7 @@ class Attachment < OpenAI::BaseModel # # @return [Array, nil] optional :tools, - -> { OpenAI::ArrayOf[union: OpenAI::Models::Beta::Threads::MessageCreateParams::Attachment::Tool] } + -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::Threads::MessageCreateParams::Attachment::Tool] } # @!parse # # @return [Array] @@ -122,10 +122,10 @@ class Attachment < OpenAI::BaseModel # # # def initialize(file_id: nil, tools: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void module Tool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type @@ -134,7 +134,7 @@ module Tool variant :file_search, -> { OpenAI::Models::Beta::Threads::MessageCreateParams::Attachment::Tool::FileSearch } - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # @!attribute type # The type of tool being defined: `file_search` # @@ -146,7 +146,7 @@ class FileSearch < OpenAI::BaseModel # # # def initialize(type: :file_search, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @!parse diff --git a/lib/openai/models/beta/threads/message_delete_params.rb b/lib/openai/models/beta/threads/message_delete_params.rb index 64604e89..759a3108 100644 --- a/lib/openai/models/beta/threads/message_delete_params.rb +++ b/lib/openai/models/beta/threads/message_delete_params.rb @@ -5,7 +5,7 @@ module Models module Beta module Threads # @see OpenAI::Resources::Beta::Threads::Messages#delete - class MessageDeleteParams < OpenAI::BaseModel + class MessageDeleteParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -21,7 +21,7 @@ class MessageDeleteParams < OpenAI::BaseModel # # # def initialize(thread_id:, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/message_deleted.rb b/lib/openai/models/beta/threads/message_deleted.rb index 488ddb42..da791175 100644 --- a/lib/openai/models/beta/threads/message_deleted.rb +++ b/lib/openai/models/beta/threads/message_deleted.rb @@ -5,7 +5,7 @@ module Models module Beta module Threads # @see OpenAI::Resources::Beta::Threads::Messages#delete - class MessageDeleted < OpenAI::BaseModel + class MessageDeleted < OpenAI::Internal::Type::BaseModel # @!attribute id # # @return [String] @@ -14,7 +14,7 @@ class MessageDeleted < OpenAI::BaseModel # @!attribute deleted # # @return [Boolean] - required :deleted, OpenAI::BooleanModel + required :deleted, OpenAI::Internal::Type::BooleanModel # @!attribute object # @@ -28,7 +28,7 @@ class MessageDeleted < OpenAI::BaseModel # # # def initialize(id:, deleted:, object: :"thread.message.deleted", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/message_delta.rb b/lib/openai/models/beta/threads/message_delta.rb index c13c0e39..b63f29bf 100644 --- a/lib/openai/models/beta/threads/message_delta.rb +++ b/lib/openai/models/beta/threads/message_delta.rb @@ -4,12 +4,13 @@ module OpenAI module Models module Beta module Threads - class MessageDelta < OpenAI::BaseModel + class MessageDelta < OpenAI::Internal::Type::BaseModel # @!attribute [r] content # The content of the message in array of text and/or images. # # @return [Array, nil] - optional :content, -> { OpenAI::ArrayOf[union: OpenAI::Models::Beta::Threads::MessageContentDelta] } + optional :content, + -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::Threads::MessageContentDelta] } # @!parse # # @return [Array] @@ -33,13 +34,13 @@ class MessageDelta < OpenAI::BaseModel # # # def initialize(content: nil, role: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The entity that produced the message. One of `user` or `assistant`. # # @see OpenAI::Models::Beta::Threads::MessageDelta#role module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum USER = :user ASSISTANT = :assistant diff --git a/lib/openai/models/beta/threads/message_delta_event.rb b/lib/openai/models/beta/threads/message_delta_event.rb index a008085e..11ba7a44 100644 --- a/lib/openai/models/beta/threads/message_delta_event.rb +++ b/lib/openai/models/beta/threads/message_delta_event.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class MessageDeltaEvent < OpenAI::BaseModel + class MessageDeltaEvent < OpenAI::Internal::Type::BaseModel # @!attribute id # The identifier of the message, which can be referenced in API endpoints. # @@ -33,7 +33,7 @@ class MessageDeltaEvent < OpenAI::BaseModel # # # def initialize(id:, delta:, object: :"thread.message.delta", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/message_list_params.rb b/lib/openai/models/beta/threads/message_list_params.rb index 1cbbff8e..9a7d62f5 100644 --- a/lib/openai/models/beta/threads/message_list_params.rb +++ b/lib/openai/models/beta/threads/message_list_params.rb @@ -5,7 +5,7 @@ module Models module Beta module Threads # @see OpenAI::Resources::Beta::Threads::Messages#list - class MessageListParams < OpenAI::BaseModel + class MessageListParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -78,12 +78,12 @@ class MessageListParams < OpenAI::BaseModel # # # def initialize(after: nil, before: nil, limit: nil, order: nil, run_id: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Sort order by the `created_at` timestamp of the objects. `asc` for ascending # order and `desc` for descending order. module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASC = :asc DESC = :desc diff --git a/lib/openai/models/beta/threads/message_retrieve_params.rb b/lib/openai/models/beta/threads/message_retrieve_params.rb index 996cb629..e2bc129b 100644 --- a/lib/openai/models/beta/threads/message_retrieve_params.rb +++ b/lib/openai/models/beta/threads/message_retrieve_params.rb @@ -5,7 +5,7 @@ module Models module Beta module Threads # @see OpenAI::Resources::Beta::Threads::Messages#retrieve - class MessageRetrieveParams < OpenAI::BaseModel + class MessageRetrieveParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -21,7 +21,7 @@ class MessageRetrieveParams < OpenAI::BaseModel # # # def initialize(thread_id:, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/message_update_params.rb b/lib/openai/models/beta/threads/message_update_params.rb index f77a34f4..c83b973f 100644 --- a/lib/openai/models/beta/threads/message_update_params.rb +++ b/lib/openai/models/beta/threads/message_update_params.rb @@ -5,7 +5,7 @@ module Models module Beta module Threads # @see OpenAI::Resources::Beta::Threads::Messages#update - class MessageUpdateParams < OpenAI::BaseModel + class MessageUpdateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -24,7 +24,7 @@ class MessageUpdateParams < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!parse # # @param thread_id [String] @@ -33,7 +33,7 @@ class MessageUpdateParams < OpenAI::BaseModel # # # def initialize(thread_id:, metadata: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/refusal_content_block.rb b/lib/openai/models/beta/threads/refusal_content_block.rb index d174b062..58ab5010 100644 --- a/lib/openai/models/beta/threads/refusal_content_block.rb +++ b/lib/openai/models/beta/threads/refusal_content_block.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class RefusalContentBlock < OpenAI::BaseModel + class RefusalContentBlock < OpenAI::Internal::Type::BaseModel # @!attribute refusal # # @return [String] @@ -24,7 +24,7 @@ class RefusalContentBlock < OpenAI::BaseModel # # # def initialize(refusal:, type: :refusal, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/refusal_delta_block.rb b/lib/openai/models/beta/threads/refusal_delta_block.rb index e65375c1..e762e97d 100644 --- a/lib/openai/models/beta/threads/refusal_delta_block.rb +++ b/lib/openai/models/beta/threads/refusal_delta_block.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class RefusalDeltaBlock < OpenAI::BaseModel + class RefusalDeltaBlock < OpenAI::Internal::Type::BaseModel # @!attribute index # The index of the refusal part in the message. # @@ -35,7 +35,7 @@ class RefusalDeltaBlock < OpenAI::BaseModel # # # def initialize(index:, refusal: nil, type: :refusal, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/required_action_function_tool_call.rb b/lib/openai/models/beta/threads/required_action_function_tool_call.rb index d9f3e412..d79d1249 100644 --- a/lib/openai/models/beta/threads/required_action_function_tool_call.rb +++ b/lib/openai/models/beta/threads/required_action_function_tool_call.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class RequiredActionFunctionToolCall < OpenAI::BaseModel + class RequiredActionFunctionToolCall < OpenAI::Internal::Type::BaseModel # @!attribute id # The ID of the tool call. This ID must be referenced when you submit the tool # outputs in using the @@ -36,10 +36,10 @@ class RequiredActionFunctionToolCall < OpenAI::BaseModel # # # def initialize(id:, function:, type: :function, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::Threads::RequiredActionFunctionToolCall#function - class Function < OpenAI::BaseModel + class Function < OpenAI::Internal::Type::BaseModel # @!attribute arguments # The arguments that the model expects you to pass to the function. # @@ -60,7 +60,7 @@ class Function < OpenAI::BaseModel # # # def initialize(arguments:, name:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/run.rb b/lib/openai/models/beta/threads/run.rb index 9b21a4a0..8af69bee 100644 --- a/lib/openai/models/beta/threads/run.rb +++ b/lib/openai/models/beta/threads/run.rb @@ -7,7 +7,7 @@ module Threads # @see OpenAI::Resources::Beta::Threads::Runs#create # # @see OpenAI::Resources::Beta::Threads::Runs#stream_raw - class Run < OpenAI::BaseModel + class Run < OpenAI::Internal::Type::BaseModel # @!attribute id # The identifier, which can be referenced in API endpoints. # @@ -96,7 +96,7 @@ class Run < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - required :metadata, OpenAI::HashOf[String], nil?: true + required :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!attribute model # The model that the @@ -118,7 +118,7 @@ class Run < OpenAI::BaseModel # during tool use. # # @return [Boolean] - required :parallel_tool_calls, OpenAI::BooleanModel + required :parallel_tool_calls, OpenAI::Internal::Type::BooleanModel # @!attribute required_action # Details on the action required to continue the run. Will be `null` if no action @@ -191,7 +191,7 @@ class Run < OpenAI::BaseModel # this run. # # @return [Array] - required :tools, -> { OpenAI::ArrayOf[union: OpenAI::Models::Beta::AssistantTool] } + required :tools, -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::AssistantTool] } # @!attribute truncation_strategy # Controls for how a thread will be truncated prior to the run. Use this to @@ -284,10 +284,10 @@ class Run < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::Threads::Run#incomplete_details - class IncompleteDetails < OpenAI::BaseModel + class IncompleteDetails < OpenAI::Internal::Type::BaseModel # @!attribute [r] reason # The reason why the run is incomplete. This will point to which specific token # limit was reached over the course of the run. @@ -307,14 +307,14 @@ class IncompleteDetails < OpenAI::BaseModel # # # def initialize(reason: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The reason why the run is incomplete. This will point to which specific token # limit was reached over the course of the run. # # @see OpenAI::Models::Beta::Threads::Run::IncompleteDetails#reason module Reason - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum MAX_COMPLETION_TOKENS = :max_completion_tokens MAX_PROMPT_TOKENS = :max_prompt_tokens @@ -328,7 +328,7 @@ module Reason end # @see OpenAI::Models::Beta::Threads::Run#last_error - class LastError < OpenAI::BaseModel + class LastError < OpenAI::Internal::Type::BaseModel # @!attribute code # One of `server_error`, `rate_limit_exceeded`, or `invalid_prompt`. # @@ -349,13 +349,13 @@ class LastError < OpenAI::BaseModel # # # def initialize(code:, message:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # One of `server_error`, `rate_limit_exceeded`, or `invalid_prompt`. # # @see OpenAI::Models::Beta::Threads::Run::LastError#code module Code - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum SERVER_ERROR = :server_error RATE_LIMIT_EXCEEDED = :rate_limit_exceeded @@ -370,7 +370,7 @@ module Code end # @see OpenAI::Models::Beta::Threads::Run#required_action - class RequiredAction < OpenAI::BaseModel + class RequiredAction < OpenAI::Internal::Type::BaseModel # @!attribute submit_tool_outputs # Details on the tool outputs needed for this run to continue. # @@ -393,16 +393,16 @@ class RequiredAction < OpenAI::BaseModel # # # def initialize(submit_tool_outputs:, type: :submit_tool_outputs, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::Threads::Run::RequiredAction#submit_tool_outputs - class SubmitToolOutputs < OpenAI::BaseModel + class SubmitToolOutputs < OpenAI::Internal::Type::BaseModel # @!attribute tool_calls # A list of the relevant tool calls. # # @return [Array] required :tool_calls, - -> { OpenAI::ArrayOf[OpenAI::Models::Beta::Threads::RequiredActionFunctionToolCall] } + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Beta::Threads::RequiredActionFunctionToolCall] } # @!parse # # Details on the tool outputs needed for this run to continue. @@ -411,12 +411,12 @@ class SubmitToolOutputs < OpenAI::BaseModel # # # def initialize(tool_calls:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end # @see OpenAI::Models::Beta::Threads::Run#truncation_strategy - class TruncationStrategy < OpenAI::BaseModel + class TruncationStrategy < OpenAI::Internal::Type::BaseModel # @!attribute type # The truncation strategy to use for the thread. The default is `auto`. If set to # `last_messages`, the thread will be truncated to the n most recent messages in @@ -442,7 +442,7 @@ class TruncationStrategy < OpenAI::BaseModel # # # def initialize(type:, last_messages: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The truncation strategy to use for the thread. The default is `auto`. If set to # `last_messages`, the thread will be truncated to the n most recent messages in @@ -451,7 +451,7 @@ class TruncationStrategy < OpenAI::BaseModel # # @see OpenAI::Models::Beta::Threads::Run::TruncationStrategy#type module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO = :auto LAST_MESSAGES = :last_messages @@ -465,7 +465,7 @@ module Type end # @see OpenAI::Models::Beta::Threads::Run#usage - class Usage < OpenAI::BaseModel + class Usage < OpenAI::Internal::Type::BaseModel # @!attribute completion_tokens # Number of completion tokens used over the course of the run. # @@ -494,7 +494,7 @@ class Usage < OpenAI::BaseModel # # # def initialize(completion_tokens:, prompt_tokens:, total_tokens:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/run_cancel_params.rb b/lib/openai/models/beta/threads/run_cancel_params.rb index bc13b744..0e9b76a0 100644 --- a/lib/openai/models/beta/threads/run_cancel_params.rb +++ b/lib/openai/models/beta/threads/run_cancel_params.rb @@ -5,7 +5,7 @@ module Models module Beta module Threads # @see OpenAI::Resources::Beta::Threads::Runs#cancel - class RunCancelParams < OpenAI::BaseModel + class RunCancelParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -21,7 +21,7 @@ class RunCancelParams < OpenAI::BaseModel # # # def initialize(thread_id:, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/run_create_params.rb b/lib/openai/models/beta/threads/run_create_params.rb index 4b60bf0b..51a64adc 100644 --- a/lib/openai/models/beta/threads/run_create_params.rb +++ b/lib/openai/models/beta/threads/run_create_params.rb @@ -7,7 +7,7 @@ module Threads # @see OpenAI::Resources::Beta::Threads::Runs#create # # @see OpenAI::Resources::Beta::Threads::Runs#stream_raw - class RunCreateParams < OpenAI::BaseModel + class RunCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -30,7 +30,8 @@ class RunCreateParams < OpenAI::BaseModel # for more information. # # @return [Array, nil] - optional :include, -> { OpenAI::ArrayOf[enum: OpenAI::Models::Beta::Threads::Runs::RunStepInclude] } + optional :include, + -> { OpenAI::Internal::Type::ArrayOf[enum: OpenAI::Models::Beta::Threads::Runs::RunStepInclude] } # @!parse # # @return [Array] @@ -49,7 +50,7 @@ class RunCreateParams < OpenAI::BaseModel # # @return [Array, nil] optional :additional_messages, - -> { OpenAI::ArrayOf[OpenAI::Models::Beta::Threads::RunCreateParams::AdditionalMessage] }, + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Beta::Threads::RunCreateParams::AdditionalMessage] }, nil?: true # @!attribute instructions @@ -89,7 +90,7 @@ class RunCreateParams < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!attribute model # The ID of the [Model](https://platform.openai.com/docs/api-reference/models) to @@ -106,7 +107,7 @@ class RunCreateParams < OpenAI::BaseModel # during tool use. # # @return [Boolean, nil] - optional :parallel_tool_calls, OpenAI::BooleanModel + optional :parallel_tool_calls, OpenAI::Internal::Type::BooleanModel # @!parse # # @return [Boolean] @@ -173,7 +174,9 @@ class RunCreateParams < OpenAI::BaseModel # modifying the behavior on a per-run basis. # # @return [Array, nil] - optional :tools, -> { OpenAI::ArrayOf[union: OpenAI::Models::Beta::AssistantTool] }, nil?: true + optional :tools, + -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::AssistantTool] }, + nil?: true # @!attribute top_p # An alternative to sampling with temperature, called nucleus sampling, where the @@ -238,9 +241,9 @@ class RunCreateParams < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class AdditionalMessage < OpenAI::BaseModel + class AdditionalMessage < OpenAI::Internal::Type::BaseModel # @!attribute content # The text contents of the message. # @@ -264,7 +267,7 @@ class AdditionalMessage < OpenAI::BaseModel # # @return [Array, nil] optional :attachments, - -> { OpenAI::ArrayOf[OpenAI::Models::Beta::Threads::RunCreateParams::AdditionalMessage::Attachment] }, + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Beta::Threads::RunCreateParams::AdditionalMessage::Attachment] }, nil?: true # @!attribute metadata @@ -276,7 +279,7 @@ class AdditionalMessage < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!parse # # @param content [String, Array] @@ -286,13 +289,13 @@ class AdditionalMessage < OpenAI::BaseModel # # # def initialize(content:, role:, attachments: nil, metadata: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The text contents of the message. # # @see OpenAI::Models::Beta::Threads::RunCreateParams::AdditionalMessage#content module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # The text contents of the message. variant String @@ -305,7 +308,7 @@ module Content # def self.variants; end MessageContentPartParamArray = - OpenAI::ArrayOf[union: -> { OpenAI::Models::Beta::Threads::MessageContentPartParam }] + OpenAI::Internal::Type::ArrayOf[union: -> { OpenAI::Models::Beta::Threads::MessageContentPartParam }] end # The role of the entity that is creating the message. Allowed values include: @@ -317,7 +320,7 @@ module Content # # @see OpenAI::Models::Beta::Threads::RunCreateParams::AdditionalMessage#role module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum USER = :user ASSISTANT = :assistant @@ -329,7 +332,7 @@ module Role # def self.values; end end - class Attachment < OpenAI::BaseModel + class Attachment < OpenAI::Internal::Type::BaseModel # @!attribute [r] file_id # The ID of the file to attach to the message. # @@ -345,7 +348,7 @@ class Attachment < OpenAI::BaseModel # # @return [Array, nil] optional :tools, - -> { OpenAI::ArrayOf[union: OpenAI::Models::Beta::Threads::RunCreateParams::AdditionalMessage::Attachment::Tool] } + -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::Threads::RunCreateParams::AdditionalMessage::Attachment::Tool] } # @!parse # # @return [Array] @@ -357,10 +360,10 @@ class Attachment < OpenAI::BaseModel # # # def initialize(file_id: nil, tools: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void module Tool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type @@ -369,7 +372,7 @@ module Tool variant :file_search, -> { OpenAI::Models::Beta::Threads::RunCreateParams::AdditionalMessage::Attachment::Tool::FileSearch } - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # @!attribute type # The type of tool being defined: `file_search` # @@ -381,7 +384,7 @@ class FileSearch < OpenAI::BaseModel # # # def initialize(type: :file_search, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @!parse @@ -396,7 +399,7 @@ class FileSearch < OpenAI::BaseModel # model associated with the assistant. If not, the model associated with the # assistant will be used. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String @@ -408,7 +411,7 @@ module Model # def self.variants; end end - class TruncationStrategy < OpenAI::BaseModel + class TruncationStrategy < OpenAI::Internal::Type::BaseModel # @!attribute type # The truncation strategy to use for the thread. The default is `auto`. If set to # `last_messages`, the thread will be truncated to the n most recent messages in @@ -434,7 +437,7 @@ class TruncationStrategy < OpenAI::BaseModel # # # def initialize(type:, last_messages: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The truncation strategy to use for the thread. The default is `auto`. If set to # `last_messages`, the thread will be truncated to the n most recent messages in @@ -443,7 +446,7 @@ class TruncationStrategy < OpenAI::BaseModel # # @see OpenAI::Models::Beta::Threads::RunCreateParams::TruncationStrategy#type module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO = :auto LAST_MESSAGES = :last_messages diff --git a/lib/openai/models/beta/threads/run_list_params.rb b/lib/openai/models/beta/threads/run_list_params.rb index 7e05072e..f2b9f759 100644 --- a/lib/openai/models/beta/threads/run_list_params.rb +++ b/lib/openai/models/beta/threads/run_list_params.rb @@ -5,7 +5,7 @@ module Models module Beta module Threads # @see OpenAI::Resources::Beta::Threads::Runs#list - class RunListParams < OpenAI::BaseModel + class RunListParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -67,12 +67,12 @@ class RunListParams < OpenAI::BaseModel # # # def initialize(after: nil, before: nil, limit: nil, order: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Sort order by the `created_at` timestamp of the objects. `asc` for ascending # order and `desc` for descending order. module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASC = :asc DESC = :desc diff --git a/lib/openai/models/beta/threads/run_retrieve_params.rb b/lib/openai/models/beta/threads/run_retrieve_params.rb index ddca0649..ca7aaf1d 100644 --- a/lib/openai/models/beta/threads/run_retrieve_params.rb +++ b/lib/openai/models/beta/threads/run_retrieve_params.rb @@ -5,7 +5,7 @@ module Models module Beta module Threads # @see OpenAI::Resources::Beta::Threads::Runs#retrieve - class RunRetrieveParams < OpenAI::BaseModel + class RunRetrieveParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -21,7 +21,7 @@ class RunRetrieveParams < OpenAI::BaseModel # # # def initialize(thread_id:, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/run_status.rb b/lib/openai/models/beta/threads/run_status.rb index 351e323f..efb9f2b6 100644 --- a/lib/openai/models/beta/threads/run_status.rb +++ b/lib/openai/models/beta/threads/run_status.rb @@ -8,7 +8,7 @@ module Threads # `requires_action`, `cancelling`, `cancelled`, `failed`, `completed`, # `incomplete`, or `expired`. module RunStatus - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum QUEUED = :queued IN_PROGRESS = :in_progress diff --git a/lib/openai/models/beta/threads/run_submit_tool_outputs_params.rb b/lib/openai/models/beta/threads/run_submit_tool_outputs_params.rb index e3579635..85d3e4d9 100644 --- a/lib/openai/models/beta/threads/run_submit_tool_outputs_params.rb +++ b/lib/openai/models/beta/threads/run_submit_tool_outputs_params.rb @@ -7,7 +7,7 @@ module Threads # @see OpenAI::Resources::Beta::Threads::Runs#submit_tool_outputs # # @see OpenAI::Resources::Beta::Threads::Runs#submit_tool_stream_raw - class RunSubmitToolOutputsParams < OpenAI::BaseModel + class RunSubmitToolOutputsParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -22,7 +22,7 @@ class RunSubmitToolOutputsParams < OpenAI::BaseModel # # @return [Array] required :tool_outputs, - -> { OpenAI::ArrayOf[OpenAI::Models::Beta::Threads::RunSubmitToolOutputsParams::ToolOutput] } + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Beta::Threads::RunSubmitToolOutputsParams::ToolOutput] } # @!parse # # @param thread_id [String] @@ -31,9 +31,9 @@ class RunSubmitToolOutputsParams < OpenAI::BaseModel # # # def initialize(thread_id:, tool_outputs:, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class ToolOutput < OpenAI::BaseModel + class ToolOutput < OpenAI::Internal::Type::BaseModel # @!attribute [r] output # The output of the tool call to be submitted to continue the run. # @@ -61,7 +61,7 @@ class ToolOutput < OpenAI::BaseModel # # # def initialize(output: nil, tool_call_id: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/run_update_params.rb b/lib/openai/models/beta/threads/run_update_params.rb index 0b5a69c2..97c387fc 100644 --- a/lib/openai/models/beta/threads/run_update_params.rb +++ b/lib/openai/models/beta/threads/run_update_params.rb @@ -5,7 +5,7 @@ module Models module Beta module Threads # @see OpenAI::Resources::Beta::Threads::Runs#update - class RunUpdateParams < OpenAI::BaseModel + class RunUpdateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -24,7 +24,7 @@ class RunUpdateParams < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!parse # # @param thread_id [String] @@ -33,7 +33,7 @@ class RunUpdateParams < OpenAI::BaseModel # # # def initialize(thread_id:, metadata: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/runs/code_interpreter_logs.rb b/lib/openai/models/beta/threads/runs/code_interpreter_logs.rb index db4bce17..9d329390 100644 --- a/lib/openai/models/beta/threads/runs/code_interpreter_logs.rb +++ b/lib/openai/models/beta/threads/runs/code_interpreter_logs.rb @@ -5,7 +5,7 @@ module Models module Beta module Threads module Runs - class CodeInterpreterLogs < OpenAI::BaseModel + class CodeInterpreterLogs < OpenAI::Internal::Type::BaseModel # @!attribute index # The index of the output in the outputs array. # @@ -37,7 +37,7 @@ class CodeInterpreterLogs < OpenAI::BaseModel # # # def initialize(index:, logs: nil, type: :logs, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/runs/code_interpreter_output_image.rb b/lib/openai/models/beta/threads/runs/code_interpreter_output_image.rb index 42262679..1fbdc55b 100644 --- a/lib/openai/models/beta/threads/runs/code_interpreter_output_image.rb +++ b/lib/openai/models/beta/threads/runs/code_interpreter_output_image.rb @@ -5,7 +5,7 @@ module Models module Beta module Threads module Runs - class CodeInterpreterOutputImage < OpenAI::BaseModel + class CodeInterpreterOutputImage < OpenAI::Internal::Type::BaseModel # @!attribute index # The index of the output in the outputs array. # @@ -34,10 +34,10 @@ class CodeInterpreterOutputImage < OpenAI::BaseModel # # # def initialize(index:, image: nil, type: :image, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::Threads::Runs::CodeInterpreterOutputImage#image - class Image < OpenAI::BaseModel + class Image < OpenAI::Internal::Type::BaseModel # @!attribute [r] file_id # The [file](https://platform.openai.com/docs/api-reference/files) ID of the # image. @@ -54,7 +54,7 @@ class Image < OpenAI::BaseModel # # # def initialize(file_id: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/runs/code_interpreter_tool_call.rb b/lib/openai/models/beta/threads/runs/code_interpreter_tool_call.rb index 62aeaf0c..4daf4c57 100644 --- a/lib/openai/models/beta/threads/runs/code_interpreter_tool_call.rb +++ b/lib/openai/models/beta/threads/runs/code_interpreter_tool_call.rb @@ -5,7 +5,7 @@ module Models module Beta module Threads module Runs - class CodeInterpreterToolCall < OpenAI::BaseModel + class CodeInterpreterToolCall < OpenAI::Internal::Type::BaseModel # @!attribute id # The ID of the tool call. # @@ -35,10 +35,10 @@ class CodeInterpreterToolCall < OpenAI::BaseModel # # # def initialize(id:, code_interpreter:, type: :code_interpreter, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCall#code_interpreter - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel # @!attribute input # The input to the Code Interpreter tool call. # @@ -52,7 +52,7 @@ class CodeInterpreter < OpenAI::BaseModel # # @return [Array] required :outputs, - -> { OpenAI::ArrayOf[union: OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCall::CodeInterpreter::Output] } + -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCall::CodeInterpreter::Output] } # @!parse # # The Code Interpreter tool call definition. @@ -62,11 +62,11 @@ class CodeInterpreter < OpenAI::BaseModel # # # def initialize(input:, outputs:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Text output from the Code Interpreter tool call as part of a run step. module Output - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type @@ -77,7 +77,7 @@ module Output variant :image, -> { OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCall::CodeInterpreter::Output::Image } - class Logs < OpenAI::BaseModel + class Logs < OpenAI::Internal::Type::BaseModel # @!attribute logs # The text output from the Code Interpreter tool call. # @@ -98,10 +98,10 @@ class Logs < OpenAI::BaseModel # # # def initialize(logs:, type: :logs, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class Image < OpenAI::BaseModel + class Image < OpenAI::Internal::Type::BaseModel # @!attribute image # # @return [OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCall::CodeInterpreter::Output::Image::Image] @@ -120,10 +120,10 @@ class Image < OpenAI::BaseModel # # # def initialize(image:, type: :image, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCall::CodeInterpreter::Output::Image#image - class Image < OpenAI::BaseModel + class Image < OpenAI::Internal::Type::BaseModel # @!attribute file_id # The [file](https://platform.openai.com/docs/api-reference/files) ID of the # image. @@ -136,7 +136,7 @@ class Image < OpenAI::BaseModel # # # def initialize(file_id:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end diff --git a/lib/openai/models/beta/threads/runs/code_interpreter_tool_call_delta.rb b/lib/openai/models/beta/threads/runs/code_interpreter_tool_call_delta.rb index 5e5b0369..6573f533 100644 --- a/lib/openai/models/beta/threads/runs/code_interpreter_tool_call_delta.rb +++ b/lib/openai/models/beta/threads/runs/code_interpreter_tool_call_delta.rb @@ -5,7 +5,7 @@ module Models module Beta module Threads module Runs - class CodeInterpreterToolCallDelta < OpenAI::BaseModel + class CodeInterpreterToolCallDelta < OpenAI::Internal::Type::BaseModel # @!attribute index # The index of the tool call in the tool calls array. # @@ -50,10 +50,10 @@ class CodeInterpreterToolCallDelta < OpenAI::BaseModel # # # def initialize(index:, id: nil, code_interpreter: nil, type: :code_interpreter, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCallDelta#code_interpreter - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel # @!attribute [r] input # The input to the Code Interpreter tool call. # @@ -71,7 +71,7 @@ class CodeInterpreter < OpenAI::BaseModel # # @return [Array, nil] optional :outputs, - -> { OpenAI::ArrayOf[union: OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCallDelta::CodeInterpreter::Output] } + -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCallDelta::CodeInterpreter::Output] } # @!parse # # @return [Array] @@ -85,11 +85,11 @@ class CodeInterpreter < OpenAI::BaseModel # # # def initialize(input: nil, outputs: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Text output from the Code Interpreter tool call as part of a run step. module Output - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type diff --git a/lib/openai/models/beta/threads/runs/file_search_tool_call.rb b/lib/openai/models/beta/threads/runs/file_search_tool_call.rb index 09267fd0..c8331a7b 100644 --- a/lib/openai/models/beta/threads/runs/file_search_tool_call.rb +++ b/lib/openai/models/beta/threads/runs/file_search_tool_call.rb @@ -5,7 +5,7 @@ module Models module Beta module Threads module Runs - class FileSearchToolCall < OpenAI::BaseModel + class FileSearchToolCall < OpenAI::Internal::Type::BaseModel # @!attribute id # The ID of the tool call object. # @@ -32,10 +32,10 @@ class FileSearchToolCall < OpenAI::BaseModel # # # def initialize(id:, file_search:, type: :file_search, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall#file_search - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # @!attribute [r] ranking_options # The ranking options for the file search. # @@ -52,7 +52,7 @@ class FileSearch < OpenAI::BaseModel # # @return [Array, nil] optional :results, - -> { OpenAI::ArrayOf[OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch::Result] } + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch::Result] } # @!parse # # @return [Array] @@ -66,10 +66,10 @@ class FileSearch < OpenAI::BaseModel # # # def initialize(ranking_options: nil, results: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch#ranking_options - class RankingOptions < OpenAI::BaseModel + class RankingOptions < OpenAI::Internal::Type::BaseModel # @!attribute ranker # The ranker to use for the file search. If not specified will use the `auto` # ranker. @@ -93,14 +93,14 @@ class RankingOptions < OpenAI::BaseModel # # # def initialize(ranker:, score_threshold:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The ranker to use for the file search. If not specified will use the `auto` # ranker. # # @see OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch::RankingOptions#ranker module Ranker - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO = :auto DEFAULT_2024_08_21 = :default_2024_08_21 @@ -113,7 +113,7 @@ module Ranker end end - class Result < OpenAI::BaseModel + class Result < OpenAI::Internal::Type::BaseModel # @!attribute file_id # The ID of the file that result was found in. # @@ -139,7 +139,7 @@ class Result < OpenAI::BaseModel # # @return [Array, nil] optional :content, - -> { OpenAI::ArrayOf[OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch::Result::Content] } + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch::Result::Content] } # @!parse # # @return [Array] @@ -155,9 +155,9 @@ class Result < OpenAI::BaseModel # # # def initialize(file_id:, file_name:, score:, content: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class Content < OpenAI::BaseModel + class Content < OpenAI::Internal::Type::BaseModel # @!attribute [r] text # The text content of the file. # @@ -185,13 +185,13 @@ class Content < OpenAI::BaseModel # # # def initialize(text: nil, type: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The type of the content. # # @see OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch::Result::Content#type module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT = :text diff --git a/lib/openai/models/beta/threads/runs/file_search_tool_call_delta.rb b/lib/openai/models/beta/threads/runs/file_search_tool_call_delta.rb index 86dd6012..eb5776d0 100644 --- a/lib/openai/models/beta/threads/runs/file_search_tool_call_delta.rb +++ b/lib/openai/models/beta/threads/runs/file_search_tool_call_delta.rb @@ -5,12 +5,12 @@ module Models module Beta module Threads module Runs - class FileSearchToolCallDelta < OpenAI::BaseModel + class FileSearchToolCallDelta < OpenAI::Internal::Type::BaseModel # @!attribute file_search # For now, this is always going to be an empty object. # # @return [Object] - required :file_search, OpenAI::Unknown + required :file_search, OpenAI::Internal::Type::Unknown # @!attribute index # The index of the tool call in the tool calls array. @@ -43,7 +43,7 @@ class FileSearchToolCallDelta < OpenAI::BaseModel # # # def initialize(file_search:, index:, id: nil, type: :file_search, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/runs/function_tool_call.rb b/lib/openai/models/beta/threads/runs/function_tool_call.rb index 9d5784e9..f0677ac7 100644 --- a/lib/openai/models/beta/threads/runs/function_tool_call.rb +++ b/lib/openai/models/beta/threads/runs/function_tool_call.rb @@ -5,7 +5,7 @@ module Models module Beta module Threads module Runs - class FunctionToolCall < OpenAI::BaseModel + class FunctionToolCall < OpenAI::Internal::Type::BaseModel # @!attribute id # The ID of the tool call object. # @@ -32,10 +32,10 @@ class FunctionToolCall < OpenAI::BaseModel # # # def initialize(id:, function:, type: :function, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::Threads::Runs::FunctionToolCall#function - class Function < OpenAI::BaseModel + class Function < OpenAI::Internal::Type::BaseModel # @!attribute arguments # The arguments passed to the function. # @@ -65,7 +65,7 @@ class Function < OpenAI::BaseModel # # # def initialize(arguments:, name:, output:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/runs/function_tool_call_delta.rb b/lib/openai/models/beta/threads/runs/function_tool_call_delta.rb index d77981e5..c4bc767e 100644 --- a/lib/openai/models/beta/threads/runs/function_tool_call_delta.rb +++ b/lib/openai/models/beta/threads/runs/function_tool_call_delta.rb @@ -5,7 +5,7 @@ module Models module Beta module Threads module Runs - class FunctionToolCallDelta < OpenAI::BaseModel + class FunctionToolCallDelta < OpenAI::Internal::Type::BaseModel # @!attribute index # The index of the tool call in the tool calls array. # @@ -47,10 +47,10 @@ class FunctionToolCallDelta < OpenAI::BaseModel # # # def initialize(index:, id: nil, function: nil, type: :function, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::Threads::Runs::FunctionToolCallDelta#function - class Function < OpenAI::BaseModel + class Function < OpenAI::Internal::Type::BaseModel # @!attribute [r] arguments # The arguments passed to the function. # @@ -88,7 +88,7 @@ class Function < OpenAI::BaseModel # # # def initialize(arguments: nil, name: nil, output: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/runs/message_creation_step_details.rb b/lib/openai/models/beta/threads/runs/message_creation_step_details.rb index f7bde3db..12b7b685 100644 --- a/lib/openai/models/beta/threads/runs/message_creation_step_details.rb +++ b/lib/openai/models/beta/threads/runs/message_creation_step_details.rb @@ -5,7 +5,7 @@ module Models module Beta module Threads module Runs - class MessageCreationStepDetails < OpenAI::BaseModel + class MessageCreationStepDetails < OpenAI::Internal::Type::BaseModel # @!attribute message_creation # # @return [OpenAI::Models::Beta::Threads::Runs::MessageCreationStepDetails::MessageCreation] @@ -26,10 +26,10 @@ class MessageCreationStepDetails < OpenAI::BaseModel # # # def initialize(message_creation:, type: :message_creation, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::Threads::Runs::MessageCreationStepDetails#message_creation - class MessageCreation < OpenAI::BaseModel + class MessageCreation < OpenAI::Internal::Type::BaseModel # @!attribute message_id # The ID of the message that was created by this run step. # @@ -41,7 +41,7 @@ class MessageCreation < OpenAI::BaseModel # # # def initialize(message_id:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/runs/run_step.rb b/lib/openai/models/beta/threads/runs/run_step.rb index 21084d23..9261d7c5 100644 --- a/lib/openai/models/beta/threads/runs/run_step.rb +++ b/lib/openai/models/beta/threads/runs/run_step.rb @@ -6,7 +6,7 @@ module Beta module Threads module Runs # @see OpenAI::Resources::Beta::Threads::Runs::Steps#retrieve - class RunStep < OpenAI::BaseModel + class RunStep < OpenAI::Internal::Type::BaseModel # @!attribute id # The identifier of the run step, which can be referenced in API endpoints. # @@ -68,7 +68,7 @@ class RunStep < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - required :metadata, OpenAI::HashOf[String], nil?: true + required :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!attribute object # The object type, which is always `thread.run.step`. @@ -158,10 +158,10 @@ class RunStep < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::Threads::Runs::RunStep#last_error - class LastError < OpenAI::BaseModel + class LastError < OpenAI::Internal::Type::BaseModel # @!attribute code # One of `server_error` or `rate_limit_exceeded`. # @@ -183,13 +183,13 @@ class LastError < OpenAI::BaseModel # # # def initialize(code:, message:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # One of `server_error` or `rate_limit_exceeded`. # # @see OpenAI::Models::Beta::Threads::Runs::RunStep::LastError#code module Code - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum SERVER_ERROR = :server_error RATE_LIMIT_EXCEEDED = :rate_limit_exceeded @@ -207,7 +207,7 @@ module Code # # @see OpenAI::Models::Beta::Threads::Runs::RunStep#status module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS = :in_progress CANCELLED = :cancelled @@ -226,7 +226,7 @@ module Status # # @see OpenAI::Models::Beta::Threads::Runs::RunStep#step_details module StepDetails - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type @@ -245,7 +245,7 @@ module StepDetails # # @see OpenAI::Models::Beta::Threads::Runs::RunStep#type module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum MESSAGE_CREATION = :message_creation TOOL_CALLS = :tool_calls @@ -258,7 +258,7 @@ module Type end # @see OpenAI::Models::Beta::Threads::Runs::RunStep#usage - class Usage < OpenAI::BaseModel + class Usage < OpenAI::Internal::Type::BaseModel # @!attribute completion_tokens # Number of completion tokens used over the course of the run step. # @@ -287,7 +287,7 @@ class Usage < OpenAI::BaseModel # # # def initialize(completion_tokens:, prompt_tokens:, total_tokens:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/runs/run_step_delta.rb b/lib/openai/models/beta/threads/runs/run_step_delta.rb index 929339da..15e16864 100644 --- a/lib/openai/models/beta/threads/runs/run_step_delta.rb +++ b/lib/openai/models/beta/threads/runs/run_step_delta.rb @@ -5,7 +5,7 @@ module Models module Beta module Threads module Runs - class RunStepDelta < OpenAI::BaseModel + class RunStepDelta < OpenAI::Internal::Type::BaseModel # @!attribute [r] step_details # The details of the run step. # @@ -23,13 +23,13 @@ class RunStepDelta < OpenAI::BaseModel # # # def initialize(step_details: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The details of the run step. # # @see OpenAI::Models::Beta::Threads::Runs::RunStepDelta#step_details module StepDetails - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type diff --git a/lib/openai/models/beta/threads/runs/run_step_delta_event.rb b/lib/openai/models/beta/threads/runs/run_step_delta_event.rb index fd2e6642..027d0490 100644 --- a/lib/openai/models/beta/threads/runs/run_step_delta_event.rb +++ b/lib/openai/models/beta/threads/runs/run_step_delta_event.rb @@ -5,7 +5,7 @@ module Models module Beta module Threads module Runs - class RunStepDeltaEvent < OpenAI::BaseModel + class RunStepDeltaEvent < OpenAI::Internal::Type::BaseModel # @!attribute id # The identifier of the run step, which can be referenced in API endpoints. # @@ -34,7 +34,7 @@ class RunStepDeltaEvent < OpenAI::BaseModel # # # def initialize(id:, delta:, object: :"thread.run.step.delta", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end diff --git a/lib/openai/models/beta/threads/runs/run_step_delta_message_delta.rb b/lib/openai/models/beta/threads/runs/run_step_delta_message_delta.rb index 7acf2e7e..178763ff 100644 --- a/lib/openai/models/beta/threads/runs/run_step_delta_message_delta.rb +++ b/lib/openai/models/beta/threads/runs/run_step_delta_message_delta.rb @@ -5,7 +5,7 @@ module Models module Beta module Threads module Runs - class RunStepDeltaMessageDelta < OpenAI::BaseModel + class RunStepDeltaMessageDelta < OpenAI::Internal::Type::BaseModel # @!attribute type # Always `message_creation`. # @@ -30,10 +30,10 @@ class RunStepDeltaMessageDelta < OpenAI::BaseModel # # # def initialize(message_creation: nil, type: :message_creation, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Beta::Threads::Runs::RunStepDeltaMessageDelta#message_creation - class MessageCreation < OpenAI::BaseModel + class MessageCreation < OpenAI::Internal::Type::BaseModel # @!attribute [r] message_id # The ID of the message that was created by this run step. # @@ -49,7 +49,7 @@ class MessageCreation < OpenAI::BaseModel # # # def initialize(message_id: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/runs/run_step_include.rb b/lib/openai/models/beta/threads/runs/run_step_include.rb index f2f05bbe..4c7fe791 100644 --- a/lib/openai/models/beta/threads/runs/run_step_include.rb +++ b/lib/openai/models/beta/threads/runs/run_step_include.rb @@ -6,7 +6,7 @@ module Beta module Threads module Runs module RunStepInclude - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum STEP_DETAILS_TOOL_CALLS_FILE_SEARCH_RESULTS_CONTENT = :"step_details.tool_calls[*].file_search.results[*].content" diff --git a/lib/openai/models/beta/threads/runs/step_list_params.rb b/lib/openai/models/beta/threads/runs/step_list_params.rb index d978c19f..a7c48d03 100644 --- a/lib/openai/models/beta/threads/runs/step_list_params.rb +++ b/lib/openai/models/beta/threads/runs/step_list_params.rb @@ -6,7 +6,7 @@ module Beta module Threads module Runs # @see OpenAI::Resources::Beta::Threads::Runs::Steps#list - class StepListParams < OpenAI::BaseModel + class StepListParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -52,7 +52,8 @@ class StepListParams < OpenAI::BaseModel # for more information. # # @return [Array, nil] - optional :include, -> { OpenAI::ArrayOf[enum: OpenAI::Models::Beta::Threads::Runs::RunStepInclude] } + optional :include, + -> { OpenAI::Internal::Type::ArrayOf[enum: OpenAI::Models::Beta::Threads::Runs::RunStepInclude] } # @!parse # # @return [Array] @@ -91,12 +92,12 @@ class StepListParams < OpenAI::BaseModel # # # def initialize(thread_id:, after: nil, before: nil, include: nil, limit: nil, order: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Sort order by the `created_at` timestamp of the objects. `asc` for ascending # order and `desc` for descending order. module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASC = :asc DESC = :desc diff --git a/lib/openai/models/beta/threads/runs/step_retrieve_params.rb b/lib/openai/models/beta/threads/runs/step_retrieve_params.rb index 5349ca7f..efbf4aaf 100644 --- a/lib/openai/models/beta/threads/runs/step_retrieve_params.rb +++ b/lib/openai/models/beta/threads/runs/step_retrieve_params.rb @@ -6,7 +6,7 @@ module Beta module Threads module Runs # @see OpenAI::Resources::Beta::Threads::Runs::Steps#retrieve - class StepRetrieveParams < OpenAI::BaseModel + class StepRetrieveParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -31,7 +31,8 @@ class StepRetrieveParams < OpenAI::BaseModel # for more information. # # @return [Array, nil] - optional :include, -> { OpenAI::ArrayOf[enum: OpenAI::Models::Beta::Threads::Runs::RunStepInclude] } + optional :include, + -> { OpenAI::Internal::Type::ArrayOf[enum: OpenAI::Models::Beta::Threads::Runs::RunStepInclude] } # @!parse # # @return [Array] @@ -45,7 +46,7 @@ class StepRetrieveParams < OpenAI::BaseModel # # # def initialize(thread_id:, run_id:, include: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/runs/tool_call.rb b/lib/openai/models/beta/threads/runs/tool_call.rb index 12ba7d73..eaee8252 100644 --- a/lib/openai/models/beta/threads/runs/tool_call.rb +++ b/lib/openai/models/beta/threads/runs/tool_call.rb @@ -7,7 +7,7 @@ module Threads module Runs # Details of the Code Interpreter tool call the run step was involved in. module ToolCall - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type diff --git a/lib/openai/models/beta/threads/runs/tool_call_delta.rb b/lib/openai/models/beta/threads/runs/tool_call_delta.rb index c8c796f3..a99db2d3 100644 --- a/lib/openai/models/beta/threads/runs/tool_call_delta.rb +++ b/lib/openai/models/beta/threads/runs/tool_call_delta.rb @@ -7,7 +7,7 @@ module Threads module Runs # Details of the Code Interpreter tool call the run step was involved in. module ToolCallDelta - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type diff --git a/lib/openai/models/beta/threads/runs/tool_call_delta_object.rb b/lib/openai/models/beta/threads/runs/tool_call_delta_object.rb index b9e864cd..7f61ee05 100644 --- a/lib/openai/models/beta/threads/runs/tool_call_delta_object.rb +++ b/lib/openai/models/beta/threads/runs/tool_call_delta_object.rb @@ -5,7 +5,7 @@ module Models module Beta module Threads module Runs - class ToolCallDeltaObject < OpenAI::BaseModel + class ToolCallDeltaObject < OpenAI::Internal::Type::BaseModel # @!attribute type # Always `tool_calls`. # @@ -18,7 +18,8 @@ class ToolCallDeltaObject < OpenAI::BaseModel # `function`. # # @return [Array, nil] - optional :tool_calls, -> { OpenAI::ArrayOf[union: OpenAI::Models::Beta::Threads::Runs::ToolCallDelta] } + optional :tool_calls, + -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::Threads::Runs::ToolCallDelta] } # @!parse # # @return [Array] @@ -32,7 +33,7 @@ class ToolCallDeltaObject < OpenAI::BaseModel # # # def initialize(tool_calls: nil, type: :tool_calls, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/runs/tool_calls_step_details.rb b/lib/openai/models/beta/threads/runs/tool_calls_step_details.rb index b833e0ff..bd9aa901 100644 --- a/lib/openai/models/beta/threads/runs/tool_calls_step_details.rb +++ b/lib/openai/models/beta/threads/runs/tool_calls_step_details.rb @@ -5,14 +5,15 @@ module Models module Beta module Threads module Runs - class ToolCallsStepDetails < OpenAI::BaseModel + class ToolCallsStepDetails < OpenAI::Internal::Type::BaseModel # @!attribute tool_calls # An array of tool calls the run step was involved in. These can be associated # with one of three types of tools: `code_interpreter`, `file_search`, or # `function`. # # @return [Array] - required :tool_calls, -> { OpenAI::ArrayOf[union: OpenAI::Models::Beta::Threads::Runs::ToolCall] } + required :tool_calls, + -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::Threads::Runs::ToolCall] } # @!attribute type # Always `tool_calls`. @@ -28,7 +29,7 @@ class ToolCallsStepDetails < OpenAI::BaseModel # # # def initialize(tool_calls:, type: :tool_calls, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/text.rb b/lib/openai/models/beta/threads/text.rb index f3b930f5..35eb23bc 100644 --- a/lib/openai/models/beta/threads/text.rb +++ b/lib/openai/models/beta/threads/text.rb @@ -4,11 +4,12 @@ module OpenAI module Models module Beta module Threads - class Text < OpenAI::BaseModel + class Text < OpenAI::Internal::Type::BaseModel # @!attribute annotations # # @return [Array] - required :annotations, -> { OpenAI::ArrayOf[union: OpenAI::Models::Beta::Threads::Annotation] } + required :annotations, + -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::Threads::Annotation] } # @!attribute value # The data that makes up the text. @@ -22,7 +23,7 @@ class Text < OpenAI::BaseModel # # # def initialize(annotations:, value:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/text_content_block.rb b/lib/openai/models/beta/threads/text_content_block.rb index b70b93fc..0e90d460 100644 --- a/lib/openai/models/beta/threads/text_content_block.rb +++ b/lib/openai/models/beta/threads/text_content_block.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class TextContentBlock < OpenAI::BaseModel + class TextContentBlock < OpenAI::Internal::Type::BaseModel # @!attribute text # # @return [OpenAI::Models::Beta::Threads::Text] @@ -24,7 +24,7 @@ class TextContentBlock < OpenAI::BaseModel # # # def initialize(text:, type: :text, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/text_content_block_param.rb b/lib/openai/models/beta/threads/text_content_block_param.rb index ce067a8c..9cbf24b3 100644 --- a/lib/openai/models/beta/threads/text_content_block_param.rb +++ b/lib/openai/models/beta/threads/text_content_block_param.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class TextContentBlockParam < OpenAI::BaseModel + class TextContentBlockParam < OpenAI::Internal::Type::BaseModel # @!attribute text # Text content to be sent to the model # @@ -25,7 +25,7 @@ class TextContentBlockParam < OpenAI::BaseModel # # # def initialize(text:, type: :text, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/text_delta.rb b/lib/openai/models/beta/threads/text_delta.rb index ee2b9db4..978ecd2c 100644 --- a/lib/openai/models/beta/threads/text_delta.rb +++ b/lib/openai/models/beta/threads/text_delta.rb @@ -4,11 +4,12 @@ module OpenAI module Models module Beta module Threads - class TextDelta < OpenAI::BaseModel + class TextDelta < OpenAI::Internal::Type::BaseModel # @!attribute [r] annotations # # @return [Array, nil] - optional :annotations, -> { OpenAI::ArrayOf[union: OpenAI::Models::Beta::Threads::AnnotationDelta] } + optional :annotations, + -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::Threads::AnnotationDelta] } # @!parse # # @return [Array] @@ -30,7 +31,7 @@ class TextDelta < OpenAI::BaseModel # # # def initialize(annotations: nil, value: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/beta/threads/text_delta_block.rb b/lib/openai/models/beta/threads/text_delta_block.rb index a06f3595..7c1b9ecc 100644 --- a/lib/openai/models/beta/threads/text_delta_block.rb +++ b/lib/openai/models/beta/threads/text_delta_block.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class TextDeltaBlock < OpenAI::BaseModel + class TextDeltaBlock < OpenAI::Internal::Type::BaseModel # @!attribute index # The index of the content part in the message. # @@ -35,7 +35,7 @@ class TextDeltaBlock < OpenAI::BaseModel # # # def initialize(index:, text: nil, type: :text, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/chat/chat_completion.rb b/lib/openai/models/chat/chat_completion.rb index 5c66a291..0f9f79a6 100644 --- a/lib/openai/models/chat/chat_completion.rb +++ b/lib/openai/models/chat/chat_completion.rb @@ -6,7 +6,7 @@ module Chat # @see OpenAI::Resources::Chat::Completions#create # # @see OpenAI::Resources::Chat::Completions#stream_raw - class ChatCompletion < OpenAI::BaseModel + class ChatCompletion < OpenAI::Internal::Type::BaseModel # @!attribute id # A unique identifier for the chat completion. # @@ -18,7 +18,7 @@ class ChatCompletion < OpenAI::BaseModel # than 1. # # @return [Array] - required :choices, -> { OpenAI::ArrayOf[OpenAI::Models::Chat::ChatCompletion::Choice] } + required :choices, -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Chat::ChatCompletion::Choice] } # @!attribute created # The Unix timestamp (in seconds) of when the chat completion was created. @@ -94,9 +94,9 @@ class ChatCompletion < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class Choice < OpenAI::BaseModel + class Choice < OpenAI::Internal::Type::BaseModel # @!attribute finish_reason # The reason the model stopped generating tokens. This will be `stop` if the model # hit a natural stop point or a provided stop sequence, `length` if the maximum @@ -134,7 +134,7 @@ class Choice < OpenAI::BaseModel # # # def initialize(finish_reason:, index:, logprobs:, message:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The reason the model stopped generating tokens. This will be `stop` if the model # hit a natural stop point or a provided stop sequence, `length` if the maximum @@ -145,7 +145,7 @@ class Choice < OpenAI::BaseModel # # @see OpenAI::Models::Chat::ChatCompletion::Choice#finish_reason module FinishReason - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum STOP = :stop LENGTH = :length @@ -161,18 +161,22 @@ module FinishReason end # @see OpenAI::Models::Chat::ChatCompletion::Choice#logprobs - class Logprobs < OpenAI::BaseModel + class Logprobs < OpenAI::Internal::Type::BaseModel # @!attribute content # A list of message content tokens with log probability information. # # @return [Array, nil] - required :content, -> { OpenAI::ArrayOf[OpenAI::Models::Chat::ChatCompletionTokenLogprob] }, nil?: true + required :content, + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Chat::ChatCompletionTokenLogprob] }, + nil?: true # @!attribute refusal # A list of message refusal tokens with log probability information. # # @return [Array, nil] - required :refusal, -> { OpenAI::ArrayOf[OpenAI::Models::Chat::ChatCompletionTokenLogprob] }, nil?: true + required :refusal, + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Chat::ChatCompletionTokenLogprob] }, + nil?: true # @!parse # # Log probability information for the choice. @@ -182,7 +186,7 @@ class Logprobs < OpenAI::BaseModel # # # def initialize(content:, refusal:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end @@ -190,7 +194,7 @@ class Logprobs < OpenAI::BaseModel # # @see OpenAI::Models::Chat::ChatCompletion#service_tier module ServiceTier - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum SCALE = :scale DEFAULT = :default diff --git a/lib/openai/models/chat/chat_completion_assistant_message_param.rb b/lib/openai/models/chat/chat_completion_assistant_message_param.rb index adf7a92f..5ac2d838 100644 --- a/lib/openai/models/chat/chat_completion_assistant_message_param.rb +++ b/lib/openai/models/chat/chat_completion_assistant_message_param.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionAssistantMessageParam < OpenAI::BaseModel + class ChatCompletionAssistantMessageParam < OpenAI::Internal::Type::BaseModel # @!attribute role # The role of the messages author, in this case `assistant`. # @@ -56,7 +56,8 @@ class ChatCompletionAssistantMessageParam < OpenAI::BaseModel # The tool calls generated by the model, such as function calls. # # @return [Array, nil] - optional :tool_calls, -> { OpenAI::ArrayOf[OpenAI::Models::Chat::ChatCompletionMessageToolCall] } + optional :tool_calls, + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Chat::ChatCompletionMessageToolCall] } # @!parse # # @return [Array] @@ -86,10 +87,10 @@ class ChatCompletionAssistantMessageParam < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Chat::ChatCompletionAssistantMessageParam#audio - class Audio < OpenAI::BaseModel + class Audio < OpenAI::Internal::Type::BaseModel # @!attribute id # Unique identifier for a previous audio response from the model. # @@ -104,7 +105,7 @@ class Audio < OpenAI::BaseModel # # # def initialize(id:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # The contents of the assistant message. Required unless `tool_calls` or @@ -112,7 +113,7 @@ class Audio < OpenAI::BaseModel # # @see OpenAI::Models::Chat::ChatCompletionAssistantMessageParam#content module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # The contents of the assistant message. variant String @@ -123,7 +124,7 @@ module Content # Learn about # [text inputs](https://platform.openai.com/docs/guides/text-generation). module ArrayOfContentPart - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type @@ -142,13 +143,13 @@ module ArrayOfContentPart # def self.variants; end ArrayOfContentPartArray = - OpenAI::ArrayOf[union: -> { OpenAI::Models::Chat::ChatCompletionAssistantMessageParam::Content::ArrayOfContentPart }] + OpenAI::Internal::Type::ArrayOf[union: -> { OpenAI::Models::Chat::ChatCompletionAssistantMessageParam::Content::ArrayOfContentPart }] end # @deprecated # # @see OpenAI::Models::Chat::ChatCompletionAssistantMessageParam#function_call - class FunctionCall < OpenAI::BaseModel + class FunctionCall < OpenAI::Internal::Type::BaseModel # @!attribute arguments # The arguments to call the function with, as generated by the model in JSON # format. Note that the model does not always generate valid JSON, and may @@ -173,7 +174,7 @@ class FunctionCall < OpenAI::BaseModel # # # def initialize(arguments:, name:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/chat/chat_completion_audio.rb b/lib/openai/models/chat/chat_completion_audio.rb index 2a8b5bca..e5c8bb97 100644 --- a/lib/openai/models/chat/chat_completion_audio.rb +++ b/lib/openai/models/chat/chat_completion_audio.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionAudio < OpenAI::BaseModel + class ChatCompletionAudio < OpenAI::Internal::Type::BaseModel # @!attribute id # Unique identifier for this audio response. # @@ -42,7 +42,7 @@ class ChatCompletionAudio < OpenAI::BaseModel # # # def initialize(id:, data:, expires_at:, transcript:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end diff --git a/lib/openai/models/chat/chat_completion_audio_param.rb b/lib/openai/models/chat/chat_completion_audio_param.rb index 9260b6ff..b1ede0ce 100644 --- a/lib/openai/models/chat/chat_completion_audio_param.rb +++ b/lib/openai/models/chat/chat_completion_audio_param.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionAudioParam < OpenAI::BaseModel + class ChatCompletionAudioParam < OpenAI::Internal::Type::BaseModel # @!attribute format_ # Specifies the output audio format. Must be one of `wav`, `mp3`, `flac`, `opus`, # or `pcm16`. @@ -28,14 +28,14 @@ class ChatCompletionAudioParam < OpenAI::BaseModel # # # def initialize(format_:, voice:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Specifies the output audio format. Must be one of `wav`, `mp3`, `flac`, `opus`, # or `pcm16`. # # @see OpenAI::Models::Chat::ChatCompletionAudioParam#format_ module Format - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum WAV = :wav MP3 = :mp3 @@ -55,7 +55,7 @@ module Format # # @see OpenAI::Models::Chat::ChatCompletionAudioParam#voice module Voice - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String diff --git a/lib/openai/models/chat/chat_completion_chunk.rb b/lib/openai/models/chat/chat_completion_chunk.rb index 3e09badb..5c693f05 100644 --- a/lib/openai/models/chat/chat_completion_chunk.rb +++ b/lib/openai/models/chat/chat_completion_chunk.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionChunk < OpenAI::BaseModel + class ChatCompletionChunk < OpenAI::Internal::Type::BaseModel # @!attribute id # A unique identifier for the chat completion. Each chunk has the same ID. # @@ -16,7 +16,8 @@ class ChatCompletionChunk < OpenAI::BaseModel # `stream_options: {"include_usage": true}`. # # @return [Array] - required :choices, -> { OpenAI::ArrayOf[OpenAI::Models::Chat::ChatCompletionChunk::Choice] } + required :choices, + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Chat::ChatCompletionChunk::Choice] } # @!attribute created # The Unix timestamp (in seconds) of when the chat completion was created. Each @@ -95,9 +96,9 @@ class ChatCompletionChunk < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class Choice < OpenAI::BaseModel + class Choice < OpenAI::Internal::Type::BaseModel # @!attribute delta # A chat completion delta generated by streamed model responses. # @@ -137,10 +138,10 @@ class Choice < OpenAI::BaseModel # # # def initialize(delta:, finish_reason:, index:, logprobs: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Chat::ChatCompletionChunk::Choice#delta - class Delta < OpenAI::BaseModel + class Delta < OpenAI::Internal::Type::BaseModel # @!attribute content # The contents of the chunk message. # @@ -178,7 +179,7 @@ class Delta < OpenAI::BaseModel # # @return [Array, nil] optional :tool_calls, - -> { OpenAI::ArrayOf[OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta::ToolCall] } + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta::ToolCall] } # @!parse # # @return [Array] @@ -195,12 +196,12 @@ class Delta < OpenAI::BaseModel # # # def initialize(content: nil, function_call: nil, refusal: nil, role: nil, tool_calls: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @deprecated # # @see OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta#function_call - class FunctionCall < OpenAI::BaseModel + class FunctionCall < OpenAI::Internal::Type::BaseModel # @!attribute [r] arguments # The arguments to call the function with, as generated by the model in JSON # format. Note that the model does not always generate valid JSON, and may @@ -233,14 +234,14 @@ class FunctionCall < OpenAI::BaseModel # # # def initialize(arguments: nil, name: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # The role of the author of this message. # # @see OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta#role module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum DEVELOPER = :developer SYSTEM = :system @@ -255,7 +256,7 @@ module Role # def self.values; end end - class ToolCall < OpenAI::BaseModel + class ToolCall < OpenAI::Internal::Type::BaseModel # @!attribute index # # @return [Integer] @@ -298,10 +299,10 @@ class ToolCall < OpenAI::BaseModel # # # def initialize(index:, id: nil, function: nil, type: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta::ToolCall#function - class Function < OpenAI::BaseModel + class Function < OpenAI::Internal::Type::BaseModel # @!attribute [r] arguments # The arguments to call the function with, as generated by the model in JSON # format. Note that the model does not always generate valid JSON, and may @@ -331,14 +332,14 @@ class Function < OpenAI::BaseModel # # # def initialize(arguments: nil, name: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # The type of the tool. Currently, only `function` is supported. # # @see OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta::ToolCall#type module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum FUNCTION = :function @@ -360,7 +361,7 @@ module Type # # @see OpenAI::Models::Chat::ChatCompletionChunk::Choice#finish_reason module FinishReason - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum STOP = :stop LENGTH = :length @@ -376,18 +377,22 @@ module FinishReason end # @see OpenAI::Models::Chat::ChatCompletionChunk::Choice#logprobs - class Logprobs < OpenAI::BaseModel + class Logprobs < OpenAI::Internal::Type::BaseModel # @!attribute content # A list of message content tokens with log probability information. # # @return [Array, nil] - required :content, -> { OpenAI::ArrayOf[OpenAI::Models::Chat::ChatCompletionTokenLogprob] }, nil?: true + required :content, + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Chat::ChatCompletionTokenLogprob] }, + nil?: true # @!attribute refusal # A list of message refusal tokens with log probability information. # # @return [Array, nil] - required :refusal, -> { OpenAI::ArrayOf[OpenAI::Models::Chat::ChatCompletionTokenLogprob] }, nil?: true + required :refusal, + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Chat::ChatCompletionTokenLogprob] }, + nil?: true # @!parse # # Log probability information for the choice. @@ -397,7 +402,7 @@ class Logprobs < OpenAI::BaseModel # # # def initialize(content:, refusal:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end @@ -405,7 +410,7 @@ class Logprobs < OpenAI::BaseModel # # @see OpenAI::Models::Chat::ChatCompletionChunk#service_tier module ServiceTier - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum SCALE = :scale DEFAULT = :default diff --git a/lib/openai/models/chat/chat_completion_content_part.rb b/lib/openai/models/chat/chat_completion_content_part.rb index dfa175c4..30da9605 100644 --- a/lib/openai/models/chat/chat_completion_content_part.rb +++ b/lib/openai/models/chat/chat_completion_content_part.rb @@ -6,7 +6,7 @@ module Chat # Learn about # [text inputs](https://platform.openai.com/docs/guides/text-generation). module ChatCompletionContentPart - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type @@ -22,7 +22,7 @@ module ChatCompletionContentPart # Learn about [file inputs](https://platform.openai.com/docs/guides/text) for text generation. variant :file, -> { OpenAI::Models::Chat::ChatCompletionContentPart::File } - class File < OpenAI::BaseModel + class File < OpenAI::Internal::Type::BaseModel # @!attribute file # # @return [OpenAI::Models::Chat::ChatCompletionContentPart::File::File] @@ -43,10 +43,10 @@ class File < OpenAI::BaseModel # # # def initialize(file:, type: :file, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Chat::ChatCompletionContentPart::File#file - class File < OpenAI::BaseModel + class File < OpenAI::Internal::Type::BaseModel # @!attribute [r] file_data # The base64 encoded file data, used when passing the file to the model as a # string. @@ -85,7 +85,7 @@ class File < OpenAI::BaseModel # # # def initialize(file_data: nil, file_id: nil, filename: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end diff --git a/lib/openai/models/chat/chat_completion_content_part_image.rb b/lib/openai/models/chat/chat_completion_content_part_image.rb index 107b95af..04a9152b 100644 --- a/lib/openai/models/chat/chat_completion_content_part_image.rb +++ b/lib/openai/models/chat/chat_completion_content_part_image.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionContentPartImage < OpenAI::BaseModel + class ChatCompletionContentPartImage < OpenAI::Internal::Type::BaseModel # @!attribute image_url # # @return [OpenAI::Models::Chat::ChatCompletionContentPartImage::ImageURL] @@ -23,10 +23,10 @@ class ChatCompletionContentPartImage < OpenAI::BaseModel # # # def initialize(image_url:, type: :image_url, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Chat::ChatCompletionContentPartImage#image_url - class ImageURL < OpenAI::BaseModel + class ImageURL < OpenAI::Internal::Type::BaseModel # @!attribute url # Either a URL of the image or the base64 encoded image data. # @@ -50,14 +50,14 @@ class ImageURL < OpenAI::BaseModel # # # def initialize(url:, detail: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Specifies the detail level of the image. Learn more in the # [Vision guide](https://platform.openai.com/docs/guides/vision#low-or-high-fidelity-image-understanding). # # @see OpenAI::Models::Chat::ChatCompletionContentPartImage::ImageURL#detail module Detail - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO = :auto LOW = :low diff --git a/lib/openai/models/chat/chat_completion_content_part_input_audio.rb b/lib/openai/models/chat/chat_completion_content_part_input_audio.rb index 4be021ce..46b3b077 100644 --- a/lib/openai/models/chat/chat_completion_content_part_input_audio.rb +++ b/lib/openai/models/chat/chat_completion_content_part_input_audio.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionContentPartInputAudio < OpenAI::BaseModel + class ChatCompletionContentPartInputAudio < OpenAI::Internal::Type::BaseModel # @!attribute input_audio # # @return [OpenAI::Models::Chat::ChatCompletionContentPartInputAudio::InputAudio] @@ -23,10 +23,10 @@ class ChatCompletionContentPartInputAudio < OpenAI::BaseModel # # # def initialize(input_audio:, type: :input_audio, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Chat::ChatCompletionContentPartInputAudio#input_audio - class InputAudio < OpenAI::BaseModel + class InputAudio < OpenAI::Internal::Type::BaseModel # @!attribute data # Base64 encoded audio data. # @@ -47,13 +47,13 @@ class InputAudio < OpenAI::BaseModel # # # def initialize(data:, format_:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The format of the encoded audio data. Currently supports "wav" and "mp3". # # @see OpenAI::Models::Chat::ChatCompletionContentPartInputAudio::InputAudio#format_ module Format - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum WAV = :wav MP3 = :mp3 diff --git a/lib/openai/models/chat/chat_completion_content_part_refusal.rb b/lib/openai/models/chat/chat_completion_content_part_refusal.rb index 83e6e914..20e3bc5d 100644 --- a/lib/openai/models/chat/chat_completion_content_part_refusal.rb +++ b/lib/openai/models/chat/chat_completion_content_part_refusal.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionContentPartRefusal < OpenAI::BaseModel + class ChatCompletionContentPartRefusal < OpenAI::Internal::Type::BaseModel # @!attribute refusal # The refusal message generated by the model. # @@ -22,7 +22,7 @@ class ChatCompletionContentPartRefusal < OpenAI::BaseModel # # # def initialize(refusal:, type: :refusal, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end diff --git a/lib/openai/models/chat/chat_completion_content_part_text.rb b/lib/openai/models/chat/chat_completion_content_part_text.rb index 1fa4cbdb..e286368a 100644 --- a/lib/openai/models/chat/chat_completion_content_part_text.rb +++ b/lib/openai/models/chat/chat_completion_content_part_text.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionContentPartText < OpenAI::BaseModel + class ChatCompletionContentPartText < OpenAI::Internal::Type::BaseModel # @!attribute text # The text content. # @@ -25,7 +25,7 @@ class ChatCompletionContentPartText < OpenAI::BaseModel # # # def initialize(text:, type: :text, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end diff --git a/lib/openai/models/chat/chat_completion_deleted.rb b/lib/openai/models/chat/chat_completion_deleted.rb index 844e2cab..a743476a 100644 --- a/lib/openai/models/chat/chat_completion_deleted.rb +++ b/lib/openai/models/chat/chat_completion_deleted.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Chat # @see OpenAI::Resources::Chat::Completions#delete - class ChatCompletionDeleted < OpenAI::BaseModel + class ChatCompletionDeleted < OpenAI::Internal::Type::BaseModel # @!attribute id # The ID of the chat completion that was deleted. # @@ -15,7 +15,7 @@ class ChatCompletionDeleted < OpenAI::BaseModel # Whether the chat completion was deleted. # # @return [Boolean] - required :deleted, OpenAI::BooleanModel + required :deleted, OpenAI::Internal::Type::BooleanModel # @!attribute object # The type of object being deleted. @@ -30,7 +30,7 @@ class ChatCompletionDeleted < OpenAI::BaseModel # # # def initialize(id:, deleted:, object: :"chat.completion.deleted", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end diff --git a/lib/openai/models/chat/chat_completion_developer_message_param.rb b/lib/openai/models/chat/chat_completion_developer_message_param.rb index 6a5af05b..1fe2ecc1 100644 --- a/lib/openai/models/chat/chat_completion_developer_message_param.rb +++ b/lib/openai/models/chat/chat_completion_developer_message_param.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionDeveloperMessageParam < OpenAI::BaseModel + class ChatCompletionDeveloperMessageParam < OpenAI::Internal::Type::BaseModel # @!attribute content # The contents of the developer message. # @@ -38,13 +38,13 @@ class ChatCompletionDeveloperMessageParam < OpenAI::BaseModel # # # def initialize(content:, name: nil, role: :developer, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The contents of the developer message. # # @see OpenAI::Models::Chat::ChatCompletionDeveloperMessageParam#content module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # The contents of the developer message. variant String @@ -57,7 +57,7 @@ module Content # def self.variants; end ChatCompletionContentPartTextArray = - OpenAI::ArrayOf[-> { OpenAI::Models::Chat::ChatCompletionContentPartText }] + OpenAI::Internal::Type::ArrayOf[-> { OpenAI::Models::Chat::ChatCompletionContentPartText }] end end end diff --git a/lib/openai/models/chat/chat_completion_function_call_option.rb b/lib/openai/models/chat/chat_completion_function_call_option.rb index 8cc72a51..9434599e 100644 --- a/lib/openai/models/chat/chat_completion_function_call_option.rb +++ b/lib/openai/models/chat/chat_completion_function_call_option.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionFunctionCallOption < OpenAI::BaseModel + class ChatCompletionFunctionCallOption < OpenAI::Internal::Type::BaseModel # @!attribute name # The name of the function to call. # @@ -18,7 +18,7 @@ class ChatCompletionFunctionCallOption < OpenAI::BaseModel # # # def initialize(name:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end diff --git a/lib/openai/models/chat/chat_completion_function_message_param.rb b/lib/openai/models/chat/chat_completion_function_message_param.rb index 1da70875..7e00e030 100644 --- a/lib/openai/models/chat/chat_completion_function_message_param.rb +++ b/lib/openai/models/chat/chat_completion_function_message_param.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Chat # @deprecated - class ChatCompletionFunctionMessageParam < OpenAI::BaseModel + class ChatCompletionFunctionMessageParam < OpenAI::Internal::Type::BaseModel # @!attribute content # The contents of the function message. # @@ -30,7 +30,7 @@ class ChatCompletionFunctionMessageParam < OpenAI::BaseModel # # # def initialize(content:, name:, role: :function, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end diff --git a/lib/openai/models/chat/chat_completion_message.rb b/lib/openai/models/chat/chat_completion_message.rb index f153dbe8..9b1828e2 100644 --- a/lib/openai/models/chat/chat_completion_message.rb +++ b/lib/openai/models/chat/chat_completion_message.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionMessage < OpenAI::BaseModel + class ChatCompletionMessage < OpenAI::Internal::Type::BaseModel # @!attribute content # The contents of the message. # @@ -27,7 +27,8 @@ class ChatCompletionMessage < OpenAI::BaseModel # [web search tool](https://platform.openai.com/docs/guides/tools-web-search?api-mode=chat). # # @return [Array, nil] - optional :annotations, -> { OpenAI::ArrayOf[OpenAI::Models::Chat::ChatCompletionMessage::Annotation] } + optional :annotations, + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Chat::ChatCompletionMessage::Annotation] } # @!parse # # @return [Array] @@ -56,7 +57,8 @@ class ChatCompletionMessage < OpenAI::BaseModel # The tool calls generated by the model, such as function calls. # # @return [Array, nil] - optional :tool_calls, -> { OpenAI::ArrayOf[OpenAI::Models::Chat::ChatCompletionMessageToolCall] } + optional :tool_calls, + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Chat::ChatCompletionMessageToolCall] } # @!parse # # @return [Array] @@ -86,9 +88,9 @@ class ChatCompletionMessage < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class Annotation < OpenAI::BaseModel + class Annotation < OpenAI::Internal::Type::BaseModel # @!attribute type # The type of the URL citation. Always `url_citation`. # @@ -109,10 +111,10 @@ class Annotation < OpenAI::BaseModel # # # def initialize(url_citation:, type: :url_citation, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Chat::ChatCompletionMessage::Annotation#url_citation - class URLCitation < OpenAI::BaseModel + class URLCitation < OpenAI::Internal::Type::BaseModel # @!attribute end_index # The index of the last character of the URL citation in the message. # @@ -147,14 +149,14 @@ class URLCitation < OpenAI::BaseModel # # # def initialize(end_index:, start_index:, title:, url:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end # @deprecated # # @see OpenAI::Models::Chat::ChatCompletionMessage#function_call - class FunctionCall < OpenAI::BaseModel + class FunctionCall < OpenAI::Internal::Type::BaseModel # @!attribute arguments # The arguments to call the function with, as generated by the model in JSON # format. Note that the model does not always generate valid JSON, and may @@ -179,7 +181,7 @@ class FunctionCall < OpenAI::BaseModel # # # def initialize(arguments:, name:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/chat/chat_completion_message_param.rb b/lib/openai/models/chat/chat_completion_message_param.rb index d55d9826..6796ea10 100644 --- a/lib/openai/models/chat/chat_completion_message_param.rb +++ b/lib/openai/models/chat/chat_completion_message_param.rb @@ -7,7 +7,7 @@ module Chat # messages sent by the user. With o1 models and newer, `developer` messages # replace the previous `system` messages. module ChatCompletionMessageParam - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :role diff --git a/lib/openai/models/chat/chat_completion_message_tool_call.rb b/lib/openai/models/chat/chat_completion_message_tool_call.rb index e114fcfe..508bc7ca 100644 --- a/lib/openai/models/chat/chat_completion_message_tool_call.rb +++ b/lib/openai/models/chat/chat_completion_message_tool_call.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionMessageToolCall < OpenAI::BaseModel + class ChatCompletionMessageToolCall < OpenAI::Internal::Type::BaseModel # @!attribute id # The ID of the tool call. # @@ -29,10 +29,10 @@ class ChatCompletionMessageToolCall < OpenAI::BaseModel # # # def initialize(id:, function:, type: :function, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Chat::ChatCompletionMessageToolCall#function - class Function < OpenAI::BaseModel + class Function < OpenAI::Internal::Type::BaseModel # @!attribute arguments # The arguments to call the function with, as generated by the model in JSON # format. Note that the model does not always generate valid JSON, and may @@ -56,7 +56,7 @@ class Function < OpenAI::BaseModel # # # def initialize(arguments:, name:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/chat/chat_completion_modality.rb b/lib/openai/models/chat/chat_completion_modality.rb index f759271a..05e0e087 100644 --- a/lib/openai/models/chat/chat_completion_modality.rb +++ b/lib/openai/models/chat/chat_completion_modality.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Chat module ChatCompletionModality - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT = :text AUDIO = :audio diff --git a/lib/openai/models/chat/chat_completion_named_tool_choice.rb b/lib/openai/models/chat/chat_completion_named_tool_choice.rb index 6d8c1713..9058c083 100644 --- a/lib/openai/models/chat/chat_completion_named_tool_choice.rb +++ b/lib/openai/models/chat/chat_completion_named_tool_choice.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionNamedToolChoice < OpenAI::BaseModel + class ChatCompletionNamedToolChoice < OpenAI::Internal::Type::BaseModel # @!attribute function # # @return [OpenAI::Models::Chat::ChatCompletionNamedToolChoice::Function] @@ -24,10 +24,10 @@ class ChatCompletionNamedToolChoice < OpenAI::BaseModel # # # def initialize(function:, type: :function, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Chat::ChatCompletionNamedToolChoice#function - class Function < OpenAI::BaseModel + class Function < OpenAI::Internal::Type::BaseModel # @!attribute name # The name of the function to call. # @@ -39,7 +39,7 @@ class Function < OpenAI::BaseModel # # # def initialize(name:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/chat/chat_completion_prediction_content.rb b/lib/openai/models/chat/chat_completion_prediction_content.rb index 5865f161..dd3150fd 100644 --- a/lib/openai/models/chat/chat_completion_prediction_content.rb +++ b/lib/openai/models/chat/chat_completion_prediction_content.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionPredictionContent < OpenAI::BaseModel + class ChatCompletionPredictionContent < OpenAI::Internal::Type::BaseModel # @!attribute content # The content that should be matched when generating a model response. If # generated tokens would match this content, the entire model response can be @@ -28,7 +28,7 @@ class ChatCompletionPredictionContent < OpenAI::BaseModel # # # def initialize(content:, type: :content, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The content that should be matched when generating a model response. If # generated tokens would match this content, the entire model response can be @@ -36,7 +36,7 @@ class ChatCompletionPredictionContent < OpenAI::BaseModel # # @see OpenAI::Models::Chat::ChatCompletionPredictionContent#content module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # The content used for a Predicted Output. This is often the # text of a file you are regenerating with minor changes. @@ -50,7 +50,7 @@ module Content # def self.variants; end ChatCompletionContentPartTextArray = - OpenAI::ArrayOf[-> { OpenAI::Models::Chat::ChatCompletionContentPartText }] + OpenAI::Internal::Type::ArrayOf[-> { OpenAI::Models::Chat::ChatCompletionContentPartText }] end end end diff --git a/lib/openai/models/chat/chat_completion_role.rb b/lib/openai/models/chat/chat_completion_role.rb index 33673904..87bd08a8 100644 --- a/lib/openai/models/chat/chat_completion_role.rb +++ b/lib/openai/models/chat/chat_completion_role.rb @@ -5,7 +5,7 @@ module Models module Chat # The role of the author of a message module ChatCompletionRole - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum DEVELOPER = :developer SYSTEM = :system diff --git a/lib/openai/models/chat/chat_completion_store_message.rb b/lib/openai/models/chat/chat_completion_store_message.rb index 6dc8cc30..3ae2370c 100644 --- a/lib/openai/models/chat/chat_completion_store_message.rb +++ b/lib/openai/models/chat/chat_completion_store_message.rb @@ -17,7 +17,7 @@ class ChatCompletionStoreMessage < OpenAI::Models::Chat::ChatCompletionMessage # # # def initialize(id:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end diff --git a/lib/openai/models/chat/chat_completion_stream_options.rb b/lib/openai/models/chat/chat_completion_stream_options.rb index 0d66702d..e7b99c89 100644 --- a/lib/openai/models/chat/chat_completion_stream_options.rb +++ b/lib/openai/models/chat/chat_completion_stream_options.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionStreamOptions < OpenAI::BaseModel + class ChatCompletionStreamOptions < OpenAI::Internal::Type::BaseModel # @!attribute [r] include_usage # If set, an additional chunk will be streamed before the `data: [DONE]` message. # The `usage` field on this chunk shows the token usage statistics for the entire @@ -14,7 +14,7 @@ class ChatCompletionStreamOptions < OpenAI::BaseModel # chunk which contains the total token usage for the request. # # @return [Boolean, nil] - optional :include_usage, OpenAI::BooleanModel + optional :include_usage, OpenAI::Internal::Type::BooleanModel # @!parse # # @return [Boolean] @@ -27,7 +27,7 @@ class ChatCompletionStreamOptions < OpenAI::BaseModel # # # def initialize(include_usage: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end diff --git a/lib/openai/models/chat/chat_completion_system_message_param.rb b/lib/openai/models/chat/chat_completion_system_message_param.rb index 6426eaad..d1d9b2f7 100644 --- a/lib/openai/models/chat/chat_completion_system_message_param.rb +++ b/lib/openai/models/chat/chat_completion_system_message_param.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionSystemMessageParam < OpenAI::BaseModel + class ChatCompletionSystemMessageParam < OpenAI::Internal::Type::BaseModel # @!attribute content # The contents of the system message. # @@ -38,13 +38,13 @@ class ChatCompletionSystemMessageParam < OpenAI::BaseModel # # # def initialize(content:, name: nil, role: :system, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The contents of the system message. # # @see OpenAI::Models::Chat::ChatCompletionSystemMessageParam#content module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # The contents of the system message. variant String @@ -57,7 +57,7 @@ module Content # def self.variants; end ChatCompletionContentPartTextArray = - OpenAI::ArrayOf[-> { OpenAI::Models::Chat::ChatCompletionContentPartText }] + OpenAI::Internal::Type::ArrayOf[-> { OpenAI::Models::Chat::ChatCompletionContentPartText }] end end end diff --git a/lib/openai/models/chat/chat_completion_token_logprob.rb b/lib/openai/models/chat/chat_completion_token_logprob.rb index d4faf321..8b6d0019 100644 --- a/lib/openai/models/chat/chat_completion_token_logprob.rb +++ b/lib/openai/models/chat/chat_completion_token_logprob.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionTokenLogprob < OpenAI::BaseModel + class ChatCompletionTokenLogprob < OpenAI::Internal::Type::BaseModel # @!attribute token # The token. # @@ -17,7 +17,7 @@ class ChatCompletionTokenLogprob < OpenAI::BaseModel # representation. Can be `null` if there is no bytes representation for the token. # # @return [Array, nil] - required :bytes, OpenAI::ArrayOf[Integer], nil?: true + required :bytes, OpenAI::Internal::Type::ArrayOf[Integer], nil?: true # @!attribute logprob # The log probability of this token, if it is within the top 20 most likely @@ -34,7 +34,7 @@ class ChatCompletionTokenLogprob < OpenAI::BaseModel # # @return [Array] required :top_logprobs, - -> { OpenAI::ArrayOf[OpenAI::Models::Chat::ChatCompletionTokenLogprob::TopLogprob] } + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Chat::ChatCompletionTokenLogprob::TopLogprob] } # @!parse # # @param token [String] @@ -44,9 +44,9 @@ class ChatCompletionTokenLogprob < OpenAI::BaseModel # # # def initialize(token:, bytes:, logprob:, top_logprobs:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class TopLogprob < OpenAI::BaseModel + class TopLogprob < OpenAI::Internal::Type::BaseModel # @!attribute token # The token. # @@ -60,7 +60,7 @@ class TopLogprob < OpenAI::BaseModel # representation. Can be `null` if there is no bytes representation for the token. # # @return [Array, nil] - required :bytes, OpenAI::ArrayOf[Integer], nil?: true + required :bytes, OpenAI::Internal::Type::ArrayOf[Integer], nil?: true # @!attribute logprob # The log probability of this token, if it is within the top 20 most likely @@ -77,7 +77,7 @@ class TopLogprob < OpenAI::BaseModel # # # def initialize(token:, bytes:, logprob:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/chat/chat_completion_tool.rb b/lib/openai/models/chat/chat_completion_tool.rb index f027725f..8737a603 100644 --- a/lib/openai/models/chat/chat_completion_tool.rb +++ b/lib/openai/models/chat/chat_completion_tool.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionTool < OpenAI::BaseModel + class ChatCompletionTool < OpenAI::Internal::Type::BaseModel # @!attribute function # # @return [OpenAI::Models::FunctionDefinition] @@ -21,7 +21,7 @@ class ChatCompletionTool < OpenAI::BaseModel # # # def initialize(function:, type: :function, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end diff --git a/lib/openai/models/chat/chat_completion_tool_choice_option.rb b/lib/openai/models/chat/chat_completion_tool_choice_option.rb index 7b6d3ff8..bb04bbc8 100644 --- a/lib/openai/models/chat/chat_completion_tool_choice_option.rb +++ b/lib/openai/models/chat/chat_completion_tool_choice_option.rb @@ -13,7 +13,7 @@ module Chat # `none` is the default when no tools are present. `auto` is the default if tools # are present. module ChatCompletionToolChoiceOption - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # `none` means the model will not call any tool and instead generates a message. `auto` means the model can pick between generating a message or calling one or more tools. `required` means the model must call one or more tools. variant enum: -> { OpenAI::Models::Chat::ChatCompletionToolChoiceOption::Auto } @@ -25,7 +25,7 @@ module ChatCompletionToolChoiceOption # `auto` means the model can pick between generating a message or calling one or # more tools. `required` means the model must call one or more tools. module Auto - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum NONE = :none AUTO = :auto diff --git a/lib/openai/models/chat/chat_completion_tool_message_param.rb b/lib/openai/models/chat/chat_completion_tool_message_param.rb index 37d6d123..20c3a014 100644 --- a/lib/openai/models/chat/chat_completion_tool_message_param.rb +++ b/lib/openai/models/chat/chat_completion_tool_message_param.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionToolMessageParam < OpenAI::BaseModel + class ChatCompletionToolMessageParam < OpenAI::Internal::Type::BaseModel # @!attribute content # The contents of the tool message. # @@ -29,13 +29,13 @@ class ChatCompletionToolMessageParam < OpenAI::BaseModel # # # def initialize(content:, tool_call_id:, role: :tool, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The contents of the tool message. # # @see OpenAI::Models::Chat::ChatCompletionToolMessageParam#content module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # The contents of the tool message. variant String @@ -48,7 +48,7 @@ module Content # def self.variants; end ChatCompletionContentPartTextArray = - OpenAI::ArrayOf[-> { OpenAI::Models::Chat::ChatCompletionContentPartText }] + OpenAI::Internal::Type::ArrayOf[-> { OpenAI::Models::Chat::ChatCompletionContentPartText }] end end end diff --git a/lib/openai/models/chat/chat_completion_user_message_param.rb b/lib/openai/models/chat/chat_completion_user_message_param.rb index eb640b44..ba2c2dbe 100644 --- a/lib/openai/models/chat/chat_completion_user_message_param.rb +++ b/lib/openai/models/chat/chat_completion_user_message_param.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionUserMessageParam < OpenAI::BaseModel + class ChatCompletionUserMessageParam < OpenAI::Internal::Type::BaseModel # @!attribute content # The contents of the user message. # @@ -37,13 +37,13 @@ class ChatCompletionUserMessageParam < OpenAI::BaseModel # # # def initialize(content:, name: nil, role: :user, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The contents of the user message. # # @see OpenAI::Models::Chat::ChatCompletionUserMessageParam#content module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # The text contents of the message. variant String @@ -56,7 +56,7 @@ module Content # def self.variants; end ChatCompletionContentPartArray = - OpenAI::ArrayOf[union: -> { OpenAI::Models::Chat::ChatCompletionContentPart }] + OpenAI::Internal::Type::ArrayOf[union: -> { OpenAI::Models::Chat::ChatCompletionContentPart }] end end end diff --git a/lib/openai/models/chat/completion_create_params.rb b/lib/openai/models/chat/completion_create_params.rb index 6b8c53a3..792691a9 100644 --- a/lib/openai/models/chat/completion_create_params.rb +++ b/lib/openai/models/chat/completion_create_params.rb @@ -6,7 +6,7 @@ module Chat # @see OpenAI::Resources::Chat::Completions#create # # @see OpenAI::Resources::Chat::Completions#stream_raw - class CompletionCreateParams < OpenAI::BaseModel + class CompletionCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -20,7 +20,8 @@ class CompletionCreateParams < OpenAI::BaseModel # [audio](https://platform.openai.com/docs/guides/audio). # # @return [Array] - required :messages, -> { OpenAI::ArrayOf[union: OpenAI::Models::Chat::ChatCompletionMessageParam] } + required :messages, + -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Chat::ChatCompletionMessageParam] } # @!attribute model # Model ID used to generate the response, like `gpt-4o` or `o1`. OpenAI offers a @@ -77,7 +78,8 @@ class CompletionCreateParams < OpenAI::BaseModel # A list of functions the model may generate JSON inputs for. # # @return [Array, nil] - optional :functions, -> { OpenAI::ArrayOf[OpenAI::Models::Chat::CompletionCreateParams::Function] } + optional :functions, + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Chat::CompletionCreateParams::Function] } # @!parse # # @return [Array] @@ -94,7 +96,7 @@ class CompletionCreateParams < OpenAI::BaseModel # or exclusive selection of the relevant token. # # @return [Hash{Symbol=>Integer}, nil] - optional :logit_bias, OpenAI::HashOf[Integer], nil?: true + optional :logit_bias, OpenAI::Internal::Type::HashOf[Integer], nil?: true # @!attribute logprobs # Whether to return log probabilities of the output tokens or not. If true, @@ -102,7 +104,7 @@ class CompletionCreateParams < OpenAI::BaseModel # `message`. # # @return [Boolean, nil] - optional :logprobs, OpenAI::BooleanModel, nil?: true + optional :logprobs, OpenAI::Internal::Type::BooleanModel, nil?: true # @!attribute max_completion_tokens # An upper bound for the number of tokens that can be generated for a completion, @@ -133,7 +135,7 @@ class CompletionCreateParams < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!attribute modalities # Output types that you would like the model to generate. Most models are capable @@ -149,7 +151,7 @@ class CompletionCreateParams < OpenAI::BaseModel # # @return [Array, nil] optional :modalities, - -> { OpenAI::ArrayOf[enum: OpenAI::Models::Chat::CompletionCreateParams::Modality] }, + -> { OpenAI::Internal::Type::ArrayOf[enum: OpenAI::Models::Chat::CompletionCreateParams::Modality] }, nil?: true # @!attribute n @@ -166,7 +168,7 @@ class CompletionCreateParams < OpenAI::BaseModel # during tool use. # # @return [Boolean, nil] - optional :parallel_tool_calls, OpenAI::BooleanModel + optional :parallel_tool_calls, OpenAI::Internal::Type::BooleanModel # @!parse # # @return [Boolean] @@ -259,7 +261,7 @@ class CompletionCreateParams < OpenAI::BaseModel # or [evals](https://platform.openai.com/docs/guides/evals) products. # # @return [Boolean, nil] - optional :store, OpenAI::BooleanModel, nil?: true + optional :store, OpenAI::Internal::Type::BooleanModel, nil?: true # @!attribute stream_options # Options for streaming response. Only set this when you set `stream: true`. @@ -300,7 +302,7 @@ class CompletionCreateParams < OpenAI::BaseModel # for. A max of 128 functions are supported. # # @return [Array, nil] - optional :tools, -> { OpenAI::ArrayOf[OpenAI::Models::Chat::ChatCompletionTool] } + optional :tools, -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Chat::ChatCompletionTool] } # @!parse # # @return [Array] @@ -418,7 +420,7 @@ class CompletionCreateParams < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Model ID used to generate the response, like `gpt-4o` or `o1`. OpenAI offers a # wide range of models with different capabilities, performance characteristics, @@ -426,7 +428,7 @@ class CompletionCreateParams < OpenAI::BaseModel # [model guide](https://platform.openai.com/docs/models) to browse and compare # available models. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String @@ -458,7 +460,7 @@ module Model # `none` is the default when no functions are present. `auto` is the default if # functions are present. module FunctionCall - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # `none` means the model will not call a function and instead generates a message. `auto` means the model can pick between generating a message or calling a function. variant enum: -> { OpenAI::Models::Chat::CompletionCreateParams::FunctionCall::FunctionCallMode } @@ -470,7 +472,7 @@ module FunctionCall # `auto` means the model can pick between generating a message or calling a # function. module FunctionCallMode - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum NONE = :none AUTO = :auto @@ -488,7 +490,7 @@ module FunctionCallMode end # @deprecated - class Function < OpenAI::BaseModel + class Function < OpenAI::Internal::Type::BaseModel # @!attribute name # The name of the function to be called. Must be a-z, A-Z, 0-9, or contain # underscores and dashes, with a maximum length of 64. @@ -517,7 +519,7 @@ class Function < OpenAI::BaseModel # Omitting `parameters` defines a function with an empty parameter list. # # @return [Hash{Symbol=>Object}, nil] - optional :parameters, OpenAI::HashOf[OpenAI::Unknown] + optional :parameters, OpenAI::Internal::Type::HashOf[OpenAI::Internal::Type::Unknown] # @!parse # # @return [Hash{Symbol=>Object}] @@ -530,11 +532,11 @@ class Function < OpenAI::BaseModel # # # def initialize(name:, description: nil, parameters: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end module Modality - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT = :text AUDIO = :audio @@ -557,7 +559,7 @@ module Modality # ensures the message the model generates is valid JSON. Using `json_schema` is # preferred for models that support it. module ResponseFormat - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # Default response format. Used to generate text responses. variant -> { OpenAI::Models::ResponseFormatText } @@ -592,7 +594,7 @@ module ResponseFormat # When this parameter is set, the response body will include the `service_tier` # utilized. module ServiceTier - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO = :auto DEFAULT = :default @@ -607,7 +609,7 @@ module ServiceTier # Up to 4 sequences where the API will stop generating further tokens. The # returned text will not contain the stop sequence. module Stop - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String @@ -617,10 +619,10 @@ module Stop # # @return [Array(String, Array)] # def self.variants; end - StringArray = OpenAI::ArrayOf[String] + StringArray = OpenAI::Internal::Type::ArrayOf[String] end - class WebSearchOptions < OpenAI::BaseModel + class WebSearchOptions < OpenAI::Internal::Type::BaseModel # @!attribute [r] search_context_size # High level guidance for the amount of context window space to use for the # search. One of `low`, `medium`, or `high`. `medium` is the default. @@ -651,14 +653,14 @@ class WebSearchOptions < OpenAI::BaseModel # # # def initialize(search_context_size: nil, user_location: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # High level guidance for the amount of context window space to use for the # search. One of `low`, `medium`, or `high`. `medium` is the default. # # @see OpenAI::Models::Chat::CompletionCreateParams::WebSearchOptions#search_context_size module SearchContextSize - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum LOW = :low MEDIUM = :medium @@ -672,7 +674,7 @@ module SearchContextSize end # @see OpenAI::Models::Chat::CompletionCreateParams::WebSearchOptions#user_location - class UserLocation < OpenAI::BaseModel + class UserLocation < OpenAI::Internal::Type::BaseModel # @!attribute approximate # Approximate location parameters for the search. # @@ -694,10 +696,10 @@ class UserLocation < OpenAI::BaseModel # # # def initialize(approximate:, type: :approximate, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Chat::CompletionCreateParams::WebSearchOptions::UserLocation#approximate - class Approximate < OpenAI::BaseModel + class Approximate < OpenAI::Internal::Type::BaseModel # @!attribute [r] city # Free text input for the city of the user, e.g. `San Francisco`. # @@ -750,7 +752,7 @@ class Approximate < OpenAI::BaseModel # # # def initialize(city: nil, country: nil, region: nil, timezone: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/chat/completion_delete_params.rb b/lib/openai/models/chat/completion_delete_params.rb index 69c7c1aa..a7441968 100644 --- a/lib/openai/models/chat/completion_delete_params.rb +++ b/lib/openai/models/chat/completion_delete_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Chat # @see OpenAI::Resources::Chat::Completions#delete - class CompletionDeleteParams < OpenAI::BaseModel + class CompletionDeleteParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -14,7 +14,7 @@ class CompletionDeleteParams < OpenAI::BaseModel # # # def initialize(request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/chat/completion_list_params.rb b/lib/openai/models/chat/completion_list_params.rb index c369717e..5e2f39ec 100644 --- a/lib/openai/models/chat/completion_list_params.rb +++ b/lib/openai/models/chat/completion_list_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Chat # @see OpenAI::Resources::Chat::Completions#list - class CompletionListParams < OpenAI::BaseModel + class CompletionListParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -35,7 +35,7 @@ class CompletionListParams < OpenAI::BaseModel # `metadata[key1]=value1&metadata[key2]=value2` # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!attribute [r] model # The model used to generate the Chat Completions. @@ -68,12 +68,12 @@ class CompletionListParams < OpenAI::BaseModel # # # def initialize(after: nil, limit: nil, metadata: nil, model: nil, order: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Sort order for Chat Completions by timestamp. Use `asc` for ascending order or # `desc` for descending order. Defaults to `asc`. module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASC = :asc DESC = :desc diff --git a/lib/openai/models/chat/completion_retrieve_params.rb b/lib/openai/models/chat/completion_retrieve_params.rb index 42086b40..29df8685 100644 --- a/lib/openai/models/chat/completion_retrieve_params.rb +++ b/lib/openai/models/chat/completion_retrieve_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Chat # @see OpenAI::Resources::Chat::Completions#retrieve - class CompletionRetrieveParams < OpenAI::BaseModel + class CompletionRetrieveParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -14,7 +14,7 @@ class CompletionRetrieveParams < OpenAI::BaseModel # # # def initialize(request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/chat/completion_update_params.rb b/lib/openai/models/chat/completion_update_params.rb index 87539dd0..54a9a688 100644 --- a/lib/openai/models/chat/completion_update_params.rb +++ b/lib/openai/models/chat/completion_update_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Chat # @see OpenAI::Resources::Chat::Completions#update - class CompletionUpdateParams < OpenAI::BaseModel + class CompletionUpdateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -18,7 +18,7 @@ class CompletionUpdateParams < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - required :metadata, OpenAI::HashOf[String], nil?: true + required :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!parse # # @param metadata [Hash{Symbol=>String}, nil] @@ -26,7 +26,7 @@ class CompletionUpdateParams < OpenAI::BaseModel # # # def initialize(metadata:, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/chat/completions/message_list_params.rb b/lib/openai/models/chat/completions/message_list_params.rb index 271056ab..18970d11 100644 --- a/lib/openai/models/chat/completions/message_list_params.rb +++ b/lib/openai/models/chat/completions/message_list_params.rb @@ -5,7 +5,7 @@ module Models module Chat module Completions # @see OpenAI::Resources::Chat::Completions::Messages#list - class MessageListParams < OpenAI::BaseModel + class MessageListParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -49,12 +49,12 @@ class MessageListParams < OpenAI::BaseModel # # # def initialize(after: nil, limit: nil, order: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Sort order for messages by timestamp. Use `asc` for ascending order or `desc` # for descending order. Defaults to `asc`. module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASC = :asc DESC = :desc diff --git a/lib/openai/models/chat_model.rb b/lib/openai/models/chat_model.rb index 3da0931e..a3021397 100644 --- a/lib/openai/models/chat_model.rb +++ b/lib/openai/models/chat_model.rb @@ -3,7 +3,7 @@ module OpenAI module Models module ChatModel - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum O3_MINI = :"o3-mini" O3_MINI_2025_01_31 = :"o3-mini-2025-01-31" diff --git a/lib/openai/models/comparison_filter.rb b/lib/openai/models/comparison_filter.rb index 313ae677..2a1f6511 100644 --- a/lib/openai/models/comparison_filter.rb +++ b/lib/openai/models/comparison_filter.rb @@ -2,7 +2,7 @@ module OpenAI module Models - class ComparisonFilter < OpenAI::BaseModel + class ComparisonFilter < OpenAI::Internal::Type::BaseModel # @!attribute key # The key to compare against the value. # @@ -39,7 +39,7 @@ class ComparisonFilter < OpenAI::BaseModel # # # def initialize(key:, type:, value:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`. # @@ -52,7 +52,7 @@ class ComparisonFilter < OpenAI::BaseModel # # @see OpenAI::Models::ComparisonFilter#type module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum EQ = :eq NE = :ne @@ -73,13 +73,13 @@ module Type # # @see OpenAI::Models::ComparisonFilter#value module Value - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String variant Float - variant OpenAI::BooleanModel + variant OpenAI::Internal::Type::BooleanModel # @!parse # # @return [Array(String, Float, Boolean)] diff --git a/lib/openai/models/completion.rb b/lib/openai/models/completion.rb index 791a65a3..5d7fd8b0 100644 --- a/lib/openai/models/completion.rb +++ b/lib/openai/models/completion.rb @@ -5,7 +5,7 @@ module Models # @see OpenAI::Resources::Completions#create # # @see OpenAI::Resources::Completions#stream_raw - class Completion < OpenAI::BaseModel + class Completion < OpenAI::Internal::Type::BaseModel # @!attribute id # A unique identifier for the completion. # @@ -16,7 +16,7 @@ class Completion < OpenAI::BaseModel # The list of completion choices the model generated for the input prompt. # # @return [Array] - required :choices, -> { OpenAI::ArrayOf[OpenAI::Models::CompletionChoice] } + required :choices, -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::CompletionChoice] } # @!attribute created # The Unix timestamp (in seconds) of when the completion was created. @@ -73,7 +73,7 @@ class Completion < OpenAI::BaseModel # # # def initialize(id:, choices:, created:, model:, system_fingerprint: nil, usage: nil, object: :text_completion, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/completion_choice.rb b/lib/openai/models/completion_choice.rb index 66f575ee..1a2084f4 100644 --- a/lib/openai/models/completion_choice.rb +++ b/lib/openai/models/completion_choice.rb @@ -2,7 +2,7 @@ module OpenAI module Models - class CompletionChoice < OpenAI::BaseModel + class CompletionChoice < OpenAI::Internal::Type::BaseModel # @!attribute finish_reason # The reason the model stopped generating tokens. This will be `stop` if the model # hit a natural stop point or a provided stop sequence, `length` if the maximum @@ -35,7 +35,7 @@ class CompletionChoice < OpenAI::BaseModel # # # def initialize(finish_reason:, index:, logprobs:, text:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The reason the model stopped generating tokens. This will be `stop` if the model # hit a natural stop point or a provided stop sequence, `length` if the maximum @@ -44,7 +44,7 @@ class CompletionChoice < OpenAI::BaseModel # # @see OpenAI::Models::CompletionChoice#finish_reason module FinishReason - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum STOP = :stop LENGTH = :length @@ -58,11 +58,11 @@ module FinishReason end # @see OpenAI::Models::CompletionChoice#logprobs - class Logprobs < OpenAI::BaseModel + class Logprobs < OpenAI::Internal::Type::BaseModel # @!attribute [r] text_offset # # @return [Array, nil] - optional :text_offset, OpenAI::ArrayOf[Integer] + optional :text_offset, OpenAI::Internal::Type::ArrayOf[Integer] # @!parse # # @return [Array] @@ -71,7 +71,7 @@ class Logprobs < OpenAI::BaseModel # @!attribute [r] token_logprobs # # @return [Array, nil] - optional :token_logprobs, OpenAI::ArrayOf[Float] + optional :token_logprobs, OpenAI::Internal::Type::ArrayOf[Float] # @!parse # # @return [Array] @@ -80,7 +80,7 @@ class Logprobs < OpenAI::BaseModel # @!attribute [r] tokens # # @return [Array, nil] - optional :tokens, OpenAI::ArrayOf[String] + optional :tokens, OpenAI::Internal::Type::ArrayOf[String] # @!parse # # @return [Array] @@ -89,7 +89,7 @@ class Logprobs < OpenAI::BaseModel # @!attribute [r] top_logprobs # # @return [ArrayFloat}>, nil] - optional :top_logprobs, OpenAI::ArrayOf[OpenAI::HashOf[Float]] + optional :top_logprobs, OpenAI::Internal::Type::ArrayOf[OpenAI::Internal::Type::HashOf[Float]] # @!parse # # @return [ArrayFloat}>] @@ -103,7 +103,7 @@ class Logprobs < OpenAI::BaseModel # # # def initialize(text_offset: nil, token_logprobs: nil, tokens: nil, top_logprobs: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/completion_create_params.rb b/lib/openai/models/completion_create_params.rb index 993c28d2..e17ce7d4 100644 --- a/lib/openai/models/completion_create_params.rb +++ b/lib/openai/models/completion_create_params.rb @@ -5,7 +5,7 @@ module Models # @see OpenAI::Resources::Completions#create # # @see OpenAI::Resources::Completions#stream_raw - class CompletionCreateParams < OpenAI::BaseModel + class CompletionCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -49,7 +49,7 @@ class CompletionCreateParams < OpenAI::BaseModel # Echo back the prompt in addition to the completion # # @return [Boolean, nil] - optional :echo, OpenAI::BooleanModel, nil?: true + optional :echo, OpenAI::Internal::Type::BooleanModel, nil?: true # @!attribute frequency_penalty # Number between -2.0 and 2.0. Positive values penalize new tokens based on their @@ -76,7 +76,7 @@ class CompletionCreateParams < OpenAI::BaseModel # from being generated. # # @return [Hash{Symbol=>Integer}, nil] - optional :logit_bias, OpenAI::HashOf[Integer], nil?: true + optional :logit_bias, OpenAI::Internal::Type::HashOf[Integer], nil?: true # @!attribute logprobs # Include the log probabilities on the `logprobs` most likely output tokens, as @@ -229,7 +229,7 @@ class CompletionCreateParams < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # ID of the model to use. You can use the # [List models](https://platform.openai.com/docs/api-reference/models/list) API to @@ -237,7 +237,7 @@ class CompletionCreateParams < OpenAI::BaseModel # [Model overview](https://platform.openai.com/docs/models) for descriptions of # them. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String @@ -267,7 +267,7 @@ module Model # training, so if a prompt is not specified the model will generate as if from the # beginning of a new document. module Prompt - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String @@ -281,17 +281,17 @@ module Prompt # # @return [Array(String, Array, Array, Array>)] # def self.variants; end - StringArray = OpenAI::ArrayOf[String] + StringArray = OpenAI::Internal::Type::ArrayOf[String] - IntegerArray = OpenAI::ArrayOf[Integer] + IntegerArray = OpenAI::Internal::Type::ArrayOf[Integer] - ArrayOfToken2DArray = OpenAI::ArrayOf[OpenAI::ArrayOf[Integer]] + ArrayOfToken2DArray = OpenAI::Internal::Type::ArrayOf[OpenAI::Internal::Type::ArrayOf[Integer]] end # Up to 4 sequences where the API will stop generating further tokens. The # returned text will not contain the stop sequence. module Stop - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String @@ -301,7 +301,7 @@ module Stop # # @return [Array(String, Array)] # def self.variants; end - StringArray = OpenAI::ArrayOf[String] + StringArray = OpenAI::Internal::Type::ArrayOf[String] end end end diff --git a/lib/openai/models/completion_usage.rb b/lib/openai/models/completion_usage.rb index 553021f3..bdcff97d 100644 --- a/lib/openai/models/completion_usage.rb +++ b/lib/openai/models/completion_usage.rb @@ -2,7 +2,7 @@ module OpenAI module Models - class CompletionUsage < OpenAI::BaseModel + class CompletionUsage < OpenAI::Internal::Type::BaseModel # @!attribute completion_tokens # Number of tokens in the generated completion. # @@ -61,10 +61,10 @@ class CompletionUsage < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::CompletionUsage#completion_tokens_details - class CompletionTokensDetails < OpenAI::BaseModel + class CompletionTokensDetails < OpenAI::Internal::Type::BaseModel # @!attribute [r] accepted_prediction_tokens # When using Predicted Outputs, the number of tokens in the prediction that # appeared in the completion. @@ -127,11 +127,11 @@ class CompletionTokensDetails < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @see OpenAI::Models::CompletionUsage#prompt_tokens_details - class PromptTokensDetails < OpenAI::BaseModel + class PromptTokensDetails < OpenAI::Internal::Type::BaseModel # @!attribute [r] audio_tokens # Audio input tokens present in the prompt. # @@ -160,7 +160,7 @@ class PromptTokensDetails < OpenAI::BaseModel # # # def initialize(audio_tokens: nil, cached_tokens: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/compound_filter.rb b/lib/openai/models/compound_filter.rb index 0a5c34a4..fa9859e2 100644 --- a/lib/openai/models/compound_filter.rb +++ b/lib/openai/models/compound_filter.rb @@ -2,13 +2,13 @@ module OpenAI module Models - class CompoundFilter < OpenAI::BaseModel + class CompoundFilter < OpenAI::Internal::Type::BaseModel # @!attribute filters # Array of filters to combine. Items can be `ComparisonFilter` or # `CompoundFilter`. # # @return [Array] - required :filters, -> { OpenAI::ArrayOf[union: OpenAI::Models::CompoundFilter::Filter] } + required :filters, -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::CompoundFilter::Filter] } # @!attribute type # Type of operation: `and` or `or`. @@ -24,17 +24,17 @@ class CompoundFilter < OpenAI::BaseModel # # # def initialize(filters:, type:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # A filter used to compare a specified attribute key to a given value using a # defined comparison operation. module Filter - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # A filter used to compare a specified attribute key to a given value using a defined comparison operation. variant -> { OpenAI::Models::ComparisonFilter } - variant OpenAI::Unknown + variant OpenAI::Internal::Type::Unknown # @!parse # # @return [Array(OpenAI::Models::ComparisonFilter, Object)] @@ -45,7 +45,7 @@ module Filter # # @see OpenAI::Models::CompoundFilter#type module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AND = :and OR = :or diff --git a/lib/openai/models/create_embedding_response.rb b/lib/openai/models/create_embedding_response.rb index 3fbd7bc4..bde0fa74 100644 --- a/lib/openai/models/create_embedding_response.rb +++ b/lib/openai/models/create_embedding_response.rb @@ -3,12 +3,12 @@ module OpenAI module Models # @see OpenAI::Resources::Embeddings#create - class CreateEmbeddingResponse < OpenAI::BaseModel + class CreateEmbeddingResponse < OpenAI::Internal::Type::BaseModel # @!attribute data # The list of embeddings generated by the model. # # @return [Array] - required :data, -> { OpenAI::ArrayOf[OpenAI::Models::Embedding] } + required :data, -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Embedding] } # @!attribute model # The name of the model used to generate the embedding. @@ -36,10 +36,10 @@ class CreateEmbeddingResponse < OpenAI::BaseModel # # # def initialize(data:, model:, usage:, object: :list, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::CreateEmbeddingResponse#usage - class Usage < OpenAI::BaseModel + class Usage < OpenAI::Internal::Type::BaseModel # @!attribute prompt_tokens # The number of tokens used by the prompt. # @@ -60,7 +60,7 @@ class Usage < OpenAI::BaseModel # # # def initialize(prompt_tokens:, total_tokens:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/embedding.rb b/lib/openai/models/embedding.rb index 5ac4c6c9..d71f4752 100644 --- a/lib/openai/models/embedding.rb +++ b/lib/openai/models/embedding.rb @@ -2,14 +2,14 @@ module OpenAI module Models - class Embedding < OpenAI::BaseModel + class Embedding < OpenAI::Internal::Type::BaseModel # @!attribute embedding # The embedding vector, which is a list of floats. The length of vector depends on # the model as listed in the # [embedding guide](https://platform.openai.com/docs/guides/embeddings). # # @return [Array] - required :embedding, OpenAI::ArrayOf[Float] + required :embedding, OpenAI::Internal::Type::ArrayOf[Float] # @!attribute index # The index of the embedding in the list of embeddings. @@ -32,7 +32,7 @@ class Embedding < OpenAI::BaseModel # # # def initialize(embedding:, index:, object: :embedding, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/embedding_create_params.rb b/lib/openai/models/embedding_create_params.rb index 12977ee8..4705cd60 100644 --- a/lib/openai/models/embedding_create_params.rb +++ b/lib/openai/models/embedding_create_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Embeddings#create - class EmbeddingCreateParams < OpenAI::BaseModel + class EmbeddingCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -75,7 +75,7 @@ class EmbeddingCreateParams < OpenAI::BaseModel # # # def initialize(input:, model:, dimensions: nil, encoding_format: nil, user: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Input text to embed, encoded as a string or array of tokens. To embed multiple # inputs in a single request, pass an array of strings or array of token arrays. @@ -86,7 +86,7 @@ class EmbeddingCreateParams < OpenAI::BaseModel # for counting tokens. Some models may also impose a limit on total number of # tokens summed across inputs. module Input - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # The string that will be turned into an embedding. variant String @@ -104,11 +104,11 @@ module Input # # @return [Array(String, Array, Array, Array>)] # def self.variants; end - StringArray = OpenAI::ArrayOf[String] + StringArray = OpenAI::Internal::Type::ArrayOf[String] - IntegerArray = OpenAI::ArrayOf[Integer] + IntegerArray = OpenAI::Internal::Type::ArrayOf[Integer] - ArrayOfToken2DArray = OpenAI::ArrayOf[OpenAI::ArrayOf[Integer]] + ArrayOfToken2DArray = OpenAI::Internal::Type::ArrayOf[OpenAI::Internal::Type::ArrayOf[Integer]] end # ID of the model to use. You can use the @@ -117,7 +117,7 @@ module Input # [Model overview](https://platform.openai.com/docs/models) for descriptions of # them. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String @@ -132,7 +132,7 @@ module Model # The format to return the embeddings in. Can be either `float` or # [`base64`](https://pypi.org/project/pybase64/). module EncodingFormat - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum FLOAT = :float BASE64 = :base64 diff --git a/lib/openai/models/embedding_model.rb b/lib/openai/models/embedding_model.rb index e2e7df6a..33ab9d7f 100644 --- a/lib/openai/models/embedding_model.rb +++ b/lib/openai/models/embedding_model.rb @@ -3,7 +3,7 @@ module OpenAI module Models module EmbeddingModel - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT_EMBEDDING_ADA_002 = :"text-embedding-ada-002" TEXT_EMBEDDING_3_SMALL = :"text-embedding-3-small" diff --git a/lib/openai/models/error_object.rb b/lib/openai/models/error_object.rb index c6aaddc6..37031a55 100644 --- a/lib/openai/models/error_object.rb +++ b/lib/openai/models/error_object.rb @@ -2,7 +2,7 @@ module OpenAI module Models - class ErrorObject < OpenAI::BaseModel + class ErrorObject < OpenAI::Internal::Type::BaseModel # @!attribute code # # @return [String, nil] @@ -31,7 +31,7 @@ class ErrorObject < OpenAI::BaseModel # # # def initialize(code:, message:, param:, type:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/file_chunking_strategy.rb b/lib/openai/models/file_chunking_strategy.rb index 5cfb6edf..0c37933e 100644 --- a/lib/openai/models/file_chunking_strategy.rb +++ b/lib/openai/models/file_chunking_strategy.rb @@ -4,7 +4,7 @@ module OpenAI module Models # The strategy used to chunk the file. module FileChunkingStrategy - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type diff --git a/lib/openai/models/file_chunking_strategy_param.rb b/lib/openai/models/file_chunking_strategy_param.rb index 38682a15..6652fa42 100644 --- a/lib/openai/models/file_chunking_strategy_param.rb +++ b/lib/openai/models/file_chunking_strategy_param.rb @@ -5,7 +5,7 @@ module Models # The chunking strategy used to chunk the file(s). If not set, will use the `auto` # strategy. Only applicable if `file_ids` is non-empty. module FileChunkingStrategyParam - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type diff --git a/lib/openai/models/file_content_params.rb b/lib/openai/models/file_content_params.rb index dcb5cd2b..96b979af 100644 --- a/lib/openai/models/file_content_params.rb +++ b/lib/openai/models/file_content_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Files#content - class FileContentParams < OpenAI::BaseModel + class FileContentParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -13,7 +13,7 @@ class FileContentParams < OpenAI::BaseModel # # # def initialize(request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/file_create_params.rb b/lib/openai/models/file_create_params.rb index 7049b873..43683821 100644 --- a/lib/openai/models/file_create_params.rb +++ b/lib/openai/models/file_create_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Files#create - class FileCreateParams < OpenAI::BaseModel + class FileCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -30,7 +30,7 @@ class FileCreateParams < OpenAI::BaseModel # # # def initialize(file:, purpose:, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/file_delete_params.rb b/lib/openai/models/file_delete_params.rb index cba48c40..fda911ad 100644 --- a/lib/openai/models/file_delete_params.rb +++ b/lib/openai/models/file_delete_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Files#delete - class FileDeleteParams < OpenAI::BaseModel + class FileDeleteParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -13,7 +13,7 @@ class FileDeleteParams < OpenAI::BaseModel # # # def initialize(request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/file_deleted.rb b/lib/openai/models/file_deleted.rb index c9841f7e..a391a385 100644 --- a/lib/openai/models/file_deleted.rb +++ b/lib/openai/models/file_deleted.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Files#delete - class FileDeleted < OpenAI::BaseModel + class FileDeleted < OpenAI::Internal::Type::BaseModel # @!attribute id # # @return [String] @@ -12,7 +12,7 @@ class FileDeleted < OpenAI::BaseModel # @!attribute deleted # # @return [Boolean] - required :deleted, OpenAI::BooleanModel + required :deleted, OpenAI::Internal::Type::BooleanModel # @!attribute object # @@ -26,7 +26,7 @@ class FileDeleted < OpenAI::BaseModel # # # def initialize(id:, deleted:, object: :file, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/file_list_params.rb b/lib/openai/models/file_list_params.rb index 53f26749..d5a7bce4 100644 --- a/lib/openai/models/file_list_params.rb +++ b/lib/openai/models/file_list_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Files#list - class FileListParams < OpenAI::BaseModel + class FileListParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -62,12 +62,12 @@ class FileListParams < OpenAI::BaseModel # # # def initialize(after: nil, limit: nil, order: nil, purpose: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Sort order by the `created_at` timestamp of the objects. `asc` for ascending # order and `desc` for descending order. module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASC = :asc DESC = :desc diff --git a/lib/openai/models/file_object.rb b/lib/openai/models/file_object.rb index 1b6ca589..d25bcfa7 100644 --- a/lib/openai/models/file_object.rb +++ b/lib/openai/models/file_object.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Files#create - class FileObject < OpenAI::BaseModel + class FileObject < OpenAI::Internal::Type::BaseModel # @!attribute id # The file identifier, which can be referenced in the API endpoints. # @@ -98,7 +98,7 @@ class FileObject < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The intended purpose of the file. Supported values are `assistants`, # `assistants_output`, `batch`, `batch_output`, `fine-tune`, `fine-tune-results` @@ -106,7 +106,7 @@ class FileObject < OpenAI::BaseModel # # @see OpenAI::Models::FileObject#purpose module Purpose - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASSISTANTS = :assistants ASSISTANTS_OUTPUT = :assistants_output @@ -130,7 +130,7 @@ module Purpose # # @see OpenAI::Models::FileObject#status module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum UPLOADED = :uploaded PROCESSED = :processed diff --git a/lib/openai/models/file_purpose.rb b/lib/openai/models/file_purpose.rb index 73d23787..3b9a9976 100644 --- a/lib/openai/models/file_purpose.rb +++ b/lib/openai/models/file_purpose.rb @@ -7,7 +7,7 @@ module Models # fine-tuning - `vision`: Images used for vision fine-tuning - `user_data`: # Flexible file type for any purpose - `evals`: Used for eval data sets module FilePurpose - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASSISTANTS = :assistants BATCH = :batch diff --git a/lib/openai/models/file_retrieve_params.rb b/lib/openai/models/file_retrieve_params.rb index 187535fd..ec128c3d 100644 --- a/lib/openai/models/file_retrieve_params.rb +++ b/lib/openai/models/file_retrieve_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Files#retrieve - class FileRetrieveParams < OpenAI::BaseModel + class FileRetrieveParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -13,7 +13,7 @@ class FileRetrieveParams < OpenAI::BaseModel # # # def initialize(request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/fine_tuning/fine_tuning_job.rb b/lib/openai/models/fine_tuning/fine_tuning_job.rb index cce986bc..9b72f23f 100644 --- a/lib/openai/models/fine_tuning/fine_tuning_job.rb +++ b/lib/openai/models/fine_tuning/fine_tuning_job.rb @@ -4,7 +4,7 @@ module OpenAI module Models module FineTuning # @see OpenAI::Resources::FineTuning::Jobs#create - class FineTuningJob < OpenAI::BaseModel + class FineTuningJob < OpenAI::Internal::Type::BaseModel # @!attribute id # The object identifier, which can be referenced in the API endpoints. # @@ -69,7 +69,7 @@ class FineTuningJob < OpenAI::BaseModel # [Files API](https://platform.openai.com/docs/api-reference/files/retrieve-contents). # # @return [Array] - required :result_files, OpenAI::ArrayOf[String] + required :result_files, OpenAI::Internal::Type::ArrayOf[String] # @!attribute seed # The seed used for the fine-tuning job. @@ -118,7 +118,7 @@ class FineTuningJob < OpenAI::BaseModel # # @return [Array, nil] optional :integrations, - -> { OpenAI::ArrayOf[OpenAI::Models::FineTuning::FineTuningJobWandbIntegrationObject] }, + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::FineTuning::FineTuningJobWandbIntegrationObject] }, nil?: true # @!attribute metadata @@ -130,7 +130,7 @@ class FineTuningJob < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!attribute [r] method_ # The method used for fine-tuning. @@ -191,10 +191,10 @@ class FineTuningJob < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::FineTuning::FineTuningJob#error - class Error < OpenAI::BaseModel + class Error < OpenAI::Internal::Type::BaseModel # @!attribute code # A machine-readable error code. # @@ -224,11 +224,11 @@ class Error < OpenAI::BaseModel # # # def initialize(code:, message:, param:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @see OpenAI::Models::FineTuning::FineTuningJob#hyperparameters - class Hyperparameters < OpenAI::BaseModel + class Hyperparameters < OpenAI::Internal::Type::BaseModel # @!attribute [r] batch_size # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. @@ -273,14 +273,14 @@ class Hyperparameters < OpenAI::BaseModel # # # def initialize(batch_size: nil, learning_rate_multiplier: nil, n_epochs: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. # # @see OpenAI::Models::FineTuning::FineTuningJob::Hyperparameters#batch_size module BatchSize - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant const: :auto @@ -296,7 +296,7 @@ module BatchSize # # @see OpenAI::Models::FineTuning::FineTuningJob::Hyperparameters#learning_rate_multiplier module LearningRateMultiplier - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant const: :auto @@ -312,7 +312,7 @@ module LearningRateMultiplier # # @see OpenAI::Models::FineTuning::FineTuningJob::Hyperparameters#n_epochs module NEpochs - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant const: :auto @@ -329,7 +329,7 @@ module NEpochs # # @see OpenAI::Models::FineTuning::FineTuningJob#status module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum VALIDATING_FILES = :validating_files QUEUED = :queued @@ -346,7 +346,7 @@ module Status end # @see OpenAI::Models::FineTuning::FineTuningJob#method_ - class Method < OpenAI::BaseModel + class Method < OpenAI::Internal::Type::BaseModel # @!attribute [r] dpo # Configuration for the DPO fine-tuning method. # @@ -386,10 +386,10 @@ class Method < OpenAI::BaseModel # # # def initialize(dpo: nil, supervised: nil, type: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::FineTuning::FineTuningJob::Method#dpo - class Dpo < OpenAI::BaseModel + class Dpo < OpenAI::Internal::Type::BaseModel # @!attribute [r] hyperparameters # The hyperparameters used for the fine-tuning job. # @@ -407,10 +407,10 @@ class Dpo < OpenAI::BaseModel # # # def initialize(hyperparameters: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::FineTuning::FineTuningJob::Method::Dpo#hyperparameters - class Hyperparameters < OpenAI::BaseModel + class Hyperparameters < OpenAI::Internal::Type::BaseModel # @!attribute [r] batch_size # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. @@ -469,14 +469,14 @@ class Hyperparameters < OpenAI::BaseModel # # # def initialize(batch_size: nil, beta: nil, learning_rate_multiplier: nil, n_epochs: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. # # @see OpenAI::Models::FineTuning::FineTuningJob::Method::Dpo::Hyperparameters#batch_size module BatchSize - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant const: :auto @@ -492,7 +492,7 @@ module BatchSize # # @see OpenAI::Models::FineTuning::FineTuningJob::Method::Dpo::Hyperparameters#beta module Beta - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant const: :auto @@ -508,7 +508,7 @@ module Beta # # @see OpenAI::Models::FineTuning::FineTuningJob::Method::Dpo::Hyperparameters#learning_rate_multiplier module LearningRateMultiplier - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant const: :auto @@ -524,7 +524,7 @@ module LearningRateMultiplier # # @see OpenAI::Models::FineTuning::FineTuningJob::Method::Dpo::Hyperparameters#n_epochs module NEpochs - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant const: :auto @@ -538,7 +538,7 @@ module NEpochs end # @see OpenAI::Models::FineTuning::FineTuningJob::Method#supervised - class Supervised < OpenAI::BaseModel + class Supervised < OpenAI::Internal::Type::BaseModel # @!attribute [r] hyperparameters # The hyperparameters used for the fine-tuning job. # @@ -557,10 +557,10 @@ class Supervised < OpenAI::BaseModel # # # def initialize(hyperparameters: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::FineTuning::FineTuningJob::Method::Supervised#hyperparameters - class Hyperparameters < OpenAI::BaseModel + class Hyperparameters < OpenAI::Internal::Type::BaseModel # @!attribute [r] batch_size # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. @@ -606,14 +606,14 @@ class Hyperparameters < OpenAI::BaseModel # # # def initialize(batch_size: nil, learning_rate_multiplier: nil, n_epochs: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. # # @see OpenAI::Models::FineTuning::FineTuningJob::Method::Supervised::Hyperparameters#batch_size module BatchSize - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant const: :auto @@ -629,7 +629,7 @@ module BatchSize # # @see OpenAI::Models::FineTuning::FineTuningJob::Method::Supervised::Hyperparameters#learning_rate_multiplier module LearningRateMultiplier - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant const: :auto @@ -645,7 +645,7 @@ module LearningRateMultiplier # # @see OpenAI::Models::FineTuning::FineTuningJob::Method::Supervised::Hyperparameters#n_epochs module NEpochs - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant const: :auto @@ -662,7 +662,7 @@ module NEpochs # # @see OpenAI::Models::FineTuning::FineTuningJob::Method#type module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum SUPERVISED = :supervised DPO = :dpo diff --git a/lib/openai/models/fine_tuning/fine_tuning_job_event.rb b/lib/openai/models/fine_tuning/fine_tuning_job_event.rb index 60995c3b..a354bb5a 100644 --- a/lib/openai/models/fine_tuning/fine_tuning_job_event.rb +++ b/lib/openai/models/fine_tuning/fine_tuning_job_event.rb @@ -4,7 +4,7 @@ module OpenAI module Models module FineTuning # @see OpenAI::Resources::FineTuning::Jobs#list_events - class FineTuningJobEvent < OpenAI::BaseModel + class FineTuningJobEvent < OpenAI::Internal::Type::BaseModel # @!attribute id # The object identifier. # @@ -39,7 +39,7 @@ class FineTuningJobEvent < OpenAI::BaseModel # The data associated with the event. # # @return [Object, nil] - optional :data, OpenAI::Unknown + optional :data, OpenAI::Internal::Type::Unknown # @!parse # # @return [Object] @@ -68,13 +68,13 @@ class FineTuningJobEvent < OpenAI::BaseModel # # # def initialize(id:, created_at:, level:, message:, data: nil, type: nil, object: :"fine_tuning.job.event", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The log level of the event. # # @see OpenAI::Models::FineTuning::FineTuningJobEvent#level module Level - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum INFO = :info WARN = :warn @@ -91,7 +91,7 @@ module Level # # @see OpenAI::Models::FineTuning::FineTuningJobEvent#type module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum MESSAGE = :message METRICS = :metrics diff --git a/lib/openai/models/fine_tuning/fine_tuning_job_wandb_integration.rb b/lib/openai/models/fine_tuning/fine_tuning_job_wandb_integration.rb index 3d4ee69a..5b1dc8c8 100644 --- a/lib/openai/models/fine_tuning/fine_tuning_job_wandb_integration.rb +++ b/lib/openai/models/fine_tuning/fine_tuning_job_wandb_integration.rb @@ -3,7 +3,7 @@ module OpenAI module Models module FineTuning - class FineTuningJobWandbIntegration < OpenAI::BaseModel + class FineTuningJobWandbIntegration < OpenAI::Internal::Type::BaseModel # @!attribute project # The name of the project that the new run will be created under. # @@ -31,7 +31,7 @@ class FineTuningJobWandbIntegration < OpenAI::BaseModel # "openai/finetune", "openai/{base-model}", "openai/{ftjob-abcdef}". # # @return [Array, nil] - optional :tags, OpenAI::ArrayOf[String] + optional :tags, OpenAI::Internal::Type::ArrayOf[String] # @!parse # # @return [Array] @@ -50,7 +50,7 @@ class FineTuningJobWandbIntegration < OpenAI::BaseModel # # # def initialize(project:, entity: nil, name: nil, tags: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end diff --git a/lib/openai/models/fine_tuning/fine_tuning_job_wandb_integration_object.rb b/lib/openai/models/fine_tuning/fine_tuning_job_wandb_integration_object.rb index 9d1a4377..05920c94 100644 --- a/lib/openai/models/fine_tuning/fine_tuning_job_wandb_integration_object.rb +++ b/lib/openai/models/fine_tuning/fine_tuning_job_wandb_integration_object.rb @@ -3,7 +3,7 @@ module OpenAI module Models module FineTuning - class FineTuningJobWandbIntegrationObject < OpenAI::BaseModel + class FineTuningJobWandbIntegrationObject < OpenAI::Internal::Type::BaseModel # @!attribute type # The type of the integration being enabled for the fine-tuning job # @@ -25,7 +25,7 @@ class FineTuningJobWandbIntegrationObject < OpenAI::BaseModel # # # def initialize(wandb:, type: :wandb, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end diff --git a/lib/openai/models/fine_tuning/job_cancel_params.rb b/lib/openai/models/fine_tuning/job_cancel_params.rb index 465bc7c2..92eda537 100644 --- a/lib/openai/models/fine_tuning/job_cancel_params.rb +++ b/lib/openai/models/fine_tuning/job_cancel_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module FineTuning # @see OpenAI::Resources::FineTuning::Jobs#cancel - class JobCancelParams < OpenAI::BaseModel + class JobCancelParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -14,7 +14,7 @@ class JobCancelParams < OpenAI::BaseModel # # # def initialize(request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/fine_tuning/job_create_params.rb b/lib/openai/models/fine_tuning/job_create_params.rb index bd2f2f88..89dcb87d 100644 --- a/lib/openai/models/fine_tuning/job_create_params.rb +++ b/lib/openai/models/fine_tuning/job_create_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module FineTuning # @see OpenAI::Resources::FineTuning::Jobs#create - class JobCreateParams < OpenAI::BaseModel + class JobCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -54,7 +54,7 @@ class JobCreateParams < OpenAI::BaseModel # # @return [Array, nil] optional :integrations, - -> { OpenAI::ArrayOf[OpenAI::Models::FineTuning::JobCreateParams::Integration] }, + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::FineTuning::JobCreateParams::Integration] }, nil?: true # @!attribute metadata @@ -66,7 +66,7 @@ class JobCreateParams < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!attribute [r] method_ # The method used for fine-tuning. @@ -141,12 +141,12 @@ class JobCreateParams < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The name of the model to fine-tune. You can select one of the # [supported models](https://platform.openai.com/docs/guides/fine-tuning#which-models-can-be-fine-tuned). module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String @@ -173,7 +173,7 @@ module Model end # @deprecated - class Hyperparameters < OpenAI::BaseModel + class Hyperparameters < OpenAI::Internal::Type::BaseModel # @!attribute [r] batch_size # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. @@ -219,14 +219,14 @@ class Hyperparameters < OpenAI::BaseModel # # # def initialize(batch_size: nil, learning_rate_multiplier: nil, n_epochs: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. # # @see OpenAI::Models::FineTuning::JobCreateParams::Hyperparameters#batch_size module BatchSize - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant const: :auto @@ -242,7 +242,7 @@ module BatchSize # # @see OpenAI::Models::FineTuning::JobCreateParams::Hyperparameters#learning_rate_multiplier module LearningRateMultiplier - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant const: :auto @@ -258,7 +258,7 @@ module LearningRateMultiplier # # @see OpenAI::Models::FineTuning::JobCreateParams::Hyperparameters#n_epochs module NEpochs - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant const: :auto @@ -270,7 +270,7 @@ module NEpochs end end - class Integration < OpenAI::BaseModel + class Integration < OpenAI::Internal::Type::BaseModel # @!attribute type # The type of integration to enable. Currently, only "wandb" (Weights and Biases) # is supported. @@ -293,10 +293,10 @@ class Integration < OpenAI::BaseModel # # # def initialize(wandb:, type: :wandb, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::FineTuning::JobCreateParams::Integration#wandb - class Wandb < OpenAI::BaseModel + class Wandb < OpenAI::Internal::Type::BaseModel # @!attribute project # The name of the project that the new run will be created under. # @@ -324,7 +324,7 @@ class Wandb < OpenAI::BaseModel # "openai/finetune", "openai/{base-model}", "openai/{ftjob-abcdef}". # # @return [Array, nil] - optional :tags, OpenAI::ArrayOf[String] + optional :tags, OpenAI::Internal::Type::ArrayOf[String] # @!parse # # @return [Array] @@ -343,11 +343,11 @@ class Wandb < OpenAI::BaseModel # # # def initialize(project:, entity: nil, name: nil, tags: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end - class Method < OpenAI::BaseModel + class Method < OpenAI::Internal::Type::BaseModel # @!attribute [r] dpo # Configuration for the DPO fine-tuning method. # @@ -387,10 +387,10 @@ class Method < OpenAI::BaseModel # # # def initialize(dpo: nil, supervised: nil, type: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::FineTuning::JobCreateParams::Method#dpo - class Dpo < OpenAI::BaseModel + class Dpo < OpenAI::Internal::Type::BaseModel # @!attribute [r] hyperparameters # The hyperparameters used for the fine-tuning job. # @@ -409,10 +409,10 @@ class Dpo < OpenAI::BaseModel # # # def initialize(hyperparameters: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::FineTuning::JobCreateParams::Method::Dpo#hyperparameters - class Hyperparameters < OpenAI::BaseModel + class Hyperparameters < OpenAI::Internal::Type::BaseModel # @!attribute [r] batch_size # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. @@ -471,14 +471,14 @@ class Hyperparameters < OpenAI::BaseModel # # # def initialize(batch_size: nil, beta: nil, learning_rate_multiplier: nil, n_epochs: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. # # @see OpenAI::Models::FineTuning::JobCreateParams::Method::Dpo::Hyperparameters#batch_size module BatchSize - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant const: :auto @@ -494,7 +494,7 @@ module BatchSize # # @see OpenAI::Models::FineTuning::JobCreateParams::Method::Dpo::Hyperparameters#beta module Beta - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant const: :auto @@ -510,7 +510,7 @@ module Beta # # @see OpenAI::Models::FineTuning::JobCreateParams::Method::Dpo::Hyperparameters#learning_rate_multiplier module LearningRateMultiplier - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant const: :auto @@ -526,7 +526,7 @@ module LearningRateMultiplier # # @see OpenAI::Models::FineTuning::JobCreateParams::Method::Dpo::Hyperparameters#n_epochs module NEpochs - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant const: :auto @@ -540,7 +540,7 @@ module NEpochs end # @see OpenAI::Models::FineTuning::JobCreateParams::Method#supervised - class Supervised < OpenAI::BaseModel + class Supervised < OpenAI::Internal::Type::BaseModel # @!attribute [r] hyperparameters # The hyperparameters used for the fine-tuning job. # @@ -559,10 +559,10 @@ class Supervised < OpenAI::BaseModel # # # def initialize(hyperparameters: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::FineTuning::JobCreateParams::Method::Supervised#hyperparameters - class Hyperparameters < OpenAI::BaseModel + class Hyperparameters < OpenAI::Internal::Type::BaseModel # @!attribute [r] batch_size # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. @@ -608,14 +608,14 @@ class Hyperparameters < OpenAI::BaseModel # # # def initialize(batch_size: nil, learning_rate_multiplier: nil, n_epochs: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. # # @see OpenAI::Models::FineTuning::JobCreateParams::Method::Supervised::Hyperparameters#batch_size module BatchSize - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant const: :auto @@ -631,7 +631,7 @@ module BatchSize # # @see OpenAI::Models::FineTuning::JobCreateParams::Method::Supervised::Hyperparameters#learning_rate_multiplier module LearningRateMultiplier - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant const: :auto @@ -647,7 +647,7 @@ module LearningRateMultiplier # # @see OpenAI::Models::FineTuning::JobCreateParams::Method::Supervised::Hyperparameters#n_epochs module NEpochs - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant const: :auto @@ -664,7 +664,7 @@ module NEpochs # # @see OpenAI::Models::FineTuning::JobCreateParams::Method#type module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum SUPERVISED = :supervised DPO = :dpo diff --git a/lib/openai/models/fine_tuning/job_list_events_params.rb b/lib/openai/models/fine_tuning/job_list_events_params.rb index 0e9abe08..d69f7e6e 100644 --- a/lib/openai/models/fine_tuning/job_list_events_params.rb +++ b/lib/openai/models/fine_tuning/job_list_events_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module FineTuning # @see OpenAI::Resources::FineTuning::Jobs#list_events - class JobListEventsParams < OpenAI::BaseModel + class JobListEventsParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -36,7 +36,7 @@ class JobListEventsParams < OpenAI::BaseModel # # # def initialize(after: nil, limit: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/fine_tuning/job_list_params.rb b/lib/openai/models/fine_tuning/job_list_params.rb index 4922697f..e0ffb31f 100644 --- a/lib/openai/models/fine_tuning/job_list_params.rb +++ b/lib/openai/models/fine_tuning/job_list_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module FineTuning # @see OpenAI::Resources::FineTuning::Jobs#list - class JobListParams < OpenAI::BaseModel + class JobListParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -34,7 +34,7 @@ class JobListParams < OpenAI::BaseModel # Alternatively, set `metadata=null` to indicate no metadata. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!parse # # @param after [String] @@ -44,7 +44,7 @@ class JobListParams < OpenAI::BaseModel # # # def initialize(after: nil, limit: nil, metadata: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/fine_tuning/job_retrieve_params.rb b/lib/openai/models/fine_tuning/job_retrieve_params.rb index 17337029..3fa511ac 100644 --- a/lib/openai/models/fine_tuning/job_retrieve_params.rb +++ b/lib/openai/models/fine_tuning/job_retrieve_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module FineTuning # @see OpenAI::Resources::FineTuning::Jobs#retrieve - class JobRetrieveParams < OpenAI::BaseModel + class JobRetrieveParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -14,7 +14,7 @@ class JobRetrieveParams < OpenAI::BaseModel # # # def initialize(request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/fine_tuning/jobs/checkpoint_list_params.rb b/lib/openai/models/fine_tuning/jobs/checkpoint_list_params.rb index 7a7b3c86..fc91cb4e 100644 --- a/lib/openai/models/fine_tuning/jobs/checkpoint_list_params.rb +++ b/lib/openai/models/fine_tuning/jobs/checkpoint_list_params.rb @@ -5,7 +5,7 @@ module Models module FineTuning module Jobs # @see OpenAI::Resources::FineTuning::Jobs::Checkpoints#list - class CheckpointListParams < OpenAI::BaseModel + class CheckpointListParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -37,7 +37,7 @@ class CheckpointListParams < OpenAI::BaseModel # # # def initialize(after: nil, limit: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/fine_tuning/jobs/fine_tuning_job_checkpoint.rb b/lib/openai/models/fine_tuning/jobs/fine_tuning_job_checkpoint.rb index 75874793..0af58cbe 100644 --- a/lib/openai/models/fine_tuning/jobs/fine_tuning_job_checkpoint.rb +++ b/lib/openai/models/fine_tuning/jobs/fine_tuning_job_checkpoint.rb @@ -5,7 +5,7 @@ module Models module FineTuning module Jobs # @see OpenAI::Resources::FineTuning::Jobs::Checkpoints#list - class FineTuningJobCheckpoint < OpenAI::BaseModel + class FineTuningJobCheckpoint < OpenAI::Internal::Type::BaseModel # @!attribute id # The checkpoint identifier, which can be referenced in the API endpoints. # @@ -73,10 +73,10 @@ class FineTuningJobCheckpoint < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::FineTuning::Jobs::FineTuningJobCheckpoint#metrics - class Metrics < OpenAI::BaseModel + class Metrics < OpenAI::Internal::Type::BaseModel # @!attribute [r] full_valid_loss # # @return [Float, nil] @@ -164,7 +164,7 @@ class Metrics < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/function_definition.rb b/lib/openai/models/function_definition.rb index 8d805ee6..c261b034 100644 --- a/lib/openai/models/function_definition.rb +++ b/lib/openai/models/function_definition.rb @@ -2,7 +2,7 @@ module OpenAI module Models - class FunctionDefinition < OpenAI::BaseModel + class FunctionDefinition < OpenAI::Internal::Type::BaseModel # @!attribute name # The name of the function to be called. Must be a-z, A-Z, 0-9, or contain # underscores and dashes, with a maximum length of 64. @@ -31,7 +31,7 @@ class FunctionDefinition < OpenAI::BaseModel # Omitting `parameters` defines a function with an empty parameter list. # # @return [Hash{Symbol=>Object}, nil] - optional :parameters, OpenAI::HashOf[OpenAI::Unknown] + optional :parameters, OpenAI::Internal::Type::HashOf[OpenAI::Internal::Type::Unknown] # @!parse # # @return [Hash{Symbol=>Object}] @@ -45,7 +45,7 @@ class FunctionDefinition < OpenAI::BaseModel # [function calling guide](docs/guides/function-calling). # # @return [Boolean, nil] - optional :strict, OpenAI::BooleanModel, nil?: true + optional :strict, OpenAI::Internal::Type::BooleanModel, nil?: true # @!parse # # @param name [String] @@ -55,7 +55,7 @@ class FunctionDefinition < OpenAI::BaseModel # # # def initialize(name:, description: nil, parameters: nil, strict: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/function_parameters.rb b/lib/openai/models/function_parameters.rb index a443f9b9..c32b337a 100644 --- a/lib/openai/models/function_parameters.rb +++ b/lib/openai/models/function_parameters.rb @@ -2,6 +2,6 @@ module OpenAI module Models - FunctionParameters = OpenAI::HashOf[OpenAI::Unknown] + FunctionParameters = OpenAI::Internal::Type::HashOf[OpenAI::Internal::Type::Unknown] end end diff --git a/lib/openai/models/image.rb b/lib/openai/models/image.rb index c8ad72d4..e3842efd 100644 --- a/lib/openai/models/image.rb +++ b/lib/openai/models/image.rb @@ -2,7 +2,7 @@ module OpenAI module Models - class Image < OpenAI::BaseModel + class Image < OpenAI::Internal::Type::BaseModel # @!attribute [r] b64_json # The base64-encoded JSON of the generated image, if `response_format` is # `b64_json`. @@ -44,7 +44,7 @@ class Image < OpenAI::BaseModel # # # def initialize(b64_json: nil, revised_prompt: nil, url: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/image_create_variation_params.rb b/lib/openai/models/image_create_variation_params.rb index fdc94a1e..06bd032f 100644 --- a/lib/openai/models/image_create_variation_params.rb +++ b/lib/openai/models/image_create_variation_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Images#create_variation - class ImageCreateVariationParams < OpenAI::BaseModel + class ImageCreateVariationParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -69,12 +69,12 @@ class ImageCreateVariationParams < OpenAI::BaseModel # # # def initialize(image:, model: nil, n: nil, response_format: nil, size: nil, user: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The model to use for image generation. Only `dall-e-2` is supported at this # time. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String @@ -90,7 +90,7 @@ module Model # `b64_json`. URLs are only valid for 60 minutes after the image has been # generated. module ResponseFormat - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum URL = :url B64_JSON = :b64_json @@ -105,7 +105,7 @@ module ResponseFormat # The size of the generated images. Must be one of `256x256`, `512x512`, or # `1024x1024`. module Size - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum SIZE_256X256 = :"256x256" SIZE_512X512 = :"512x512" diff --git a/lib/openai/models/image_edit_params.rb b/lib/openai/models/image_edit_params.rb index f14ec178..43a5120c 100644 --- a/lib/openai/models/image_edit_params.rb +++ b/lib/openai/models/image_edit_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Images#edit - class ImageEditParams < OpenAI::BaseModel + class ImageEditParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -100,12 +100,12 @@ class ImageEditParams < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The model to use for image generation. Only `dall-e-2` is supported at this # time. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String @@ -121,7 +121,7 @@ module Model # `b64_json`. URLs are only valid for 60 minutes after the image has been # generated. module ResponseFormat - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum URL = :url B64_JSON = :b64_json @@ -136,7 +136,7 @@ module ResponseFormat # The size of the generated images. Must be one of `256x256`, `512x512`, or # `1024x1024`. module Size - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum SIZE_256X256 = :"256x256" SIZE_512X512 = :"512x512" diff --git a/lib/openai/models/image_generate_params.rb b/lib/openai/models/image_generate_params.rb index 38a73865..30a2eaa3 100644 --- a/lib/openai/models/image_generate_params.rb +++ b/lib/openai/models/image_generate_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Images#generate - class ImageGenerateParams < OpenAI::BaseModel + class ImageGenerateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -103,11 +103,11 @@ class ImageGenerateParams < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The model to use for image generation. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String @@ -123,7 +123,7 @@ module Model # details and greater consistency across the image. This param is only supported # for `dall-e-3`. module Quality - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum STANDARD = :standard HD = :hd @@ -139,7 +139,7 @@ module Quality # `b64_json`. URLs are only valid for 60 minutes after the image has been # generated. module ResponseFormat - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum URL = :url B64_JSON = :b64_json @@ -155,7 +155,7 @@ module ResponseFormat # `1024x1024` for `dall-e-2`. Must be one of `1024x1024`, `1792x1024`, or # `1024x1792` for `dall-e-3` models. module Size - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum SIZE_256X256 = :"256x256" SIZE_512X512 = :"512x512" @@ -175,7 +175,7 @@ module Size # Natural causes the model to produce more natural, less hyper-real looking # images. This param is only supported for `dall-e-3`. module Style - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum VIVID = :vivid NATURAL = :natural diff --git a/lib/openai/models/image_model.rb b/lib/openai/models/image_model.rb index 5f3d315c..83fd2f56 100644 --- a/lib/openai/models/image_model.rb +++ b/lib/openai/models/image_model.rb @@ -3,7 +3,7 @@ module OpenAI module Models module ImageModel - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum DALL_E_2 = :"dall-e-2" DALL_E_3 = :"dall-e-3" diff --git a/lib/openai/models/images_response.rb b/lib/openai/models/images_response.rb index fb015bb2..108b9120 100644 --- a/lib/openai/models/images_response.rb +++ b/lib/openai/models/images_response.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Images#create_variation - class ImagesResponse < OpenAI::BaseModel + class ImagesResponse < OpenAI::Internal::Type::BaseModel # @!attribute created # # @return [Integer] @@ -12,7 +12,7 @@ class ImagesResponse < OpenAI::BaseModel # @!attribute data # # @return [Array] - required :data, -> { OpenAI::ArrayOf[OpenAI::Models::Image] } + required :data, -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Image] } # @!parse # # @param created [Integer] @@ -20,7 +20,7 @@ class ImagesResponse < OpenAI::BaseModel # # # def initialize(created:, data:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/metadata.rb b/lib/openai/models/metadata.rb index fea222e0..c6a6935d 100644 --- a/lib/openai/models/metadata.rb +++ b/lib/openai/models/metadata.rb @@ -2,6 +2,6 @@ module OpenAI module Models - Metadata = OpenAI::HashOf[String] + Metadata = OpenAI::Internal::Type::HashOf[String] end end diff --git a/lib/openai/models/model.rb b/lib/openai/models/model.rb index 3e17e2d3..a3d362fb 100644 --- a/lib/openai/models/model.rb +++ b/lib/openai/models/model.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Models#retrieve - class Model < OpenAI::BaseModel + class Model < OpenAI::Internal::Type::BaseModel # @!attribute id # The model identifier, which can be referenced in the API endpoints. # @@ -38,7 +38,7 @@ class Model < OpenAI::BaseModel # # # def initialize(id:, created:, owned_by:, object: :model, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/model_delete_params.rb b/lib/openai/models/model_delete_params.rb index 7a3e61da..3f4036d0 100644 --- a/lib/openai/models/model_delete_params.rb +++ b/lib/openai/models/model_delete_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Models#delete - class ModelDeleteParams < OpenAI::BaseModel + class ModelDeleteParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -13,7 +13,7 @@ class ModelDeleteParams < OpenAI::BaseModel # # # def initialize(request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/model_deleted.rb b/lib/openai/models/model_deleted.rb index f269390a..b35f6781 100644 --- a/lib/openai/models/model_deleted.rb +++ b/lib/openai/models/model_deleted.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Models#delete - class ModelDeleted < OpenAI::BaseModel + class ModelDeleted < OpenAI::Internal::Type::BaseModel # @!attribute id # # @return [String] @@ -12,7 +12,7 @@ class ModelDeleted < OpenAI::BaseModel # @!attribute deleted # # @return [Boolean] - required :deleted, OpenAI::BooleanModel + required :deleted, OpenAI::Internal::Type::BooleanModel # @!attribute object # @@ -26,7 +26,7 @@ class ModelDeleted < OpenAI::BaseModel # # # def initialize(id:, deleted:, object:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/model_list_params.rb b/lib/openai/models/model_list_params.rb index eb621c9b..c7c05067 100644 --- a/lib/openai/models/model_list_params.rb +++ b/lib/openai/models/model_list_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Models#list - class ModelListParams < OpenAI::BaseModel + class ModelListParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -13,7 +13,7 @@ class ModelListParams < OpenAI::BaseModel # # # def initialize(request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/model_retrieve_params.rb b/lib/openai/models/model_retrieve_params.rb index 330c276b..b1384fc8 100644 --- a/lib/openai/models/model_retrieve_params.rb +++ b/lib/openai/models/model_retrieve_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Models#retrieve - class ModelRetrieveParams < OpenAI::BaseModel + class ModelRetrieveParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -13,7 +13,7 @@ class ModelRetrieveParams < OpenAI::BaseModel # # # def initialize(request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/moderation.rb b/lib/openai/models/moderation.rb index 9d8bf7a0..88fd634b 100644 --- a/lib/openai/models/moderation.rb +++ b/lib/openai/models/moderation.rb @@ -2,7 +2,7 @@ module OpenAI module Models - class Moderation < OpenAI::BaseModel + class Moderation < OpenAI::Internal::Type::BaseModel # @!attribute categories # A list of the categories, and whether they are flagged or not. # @@ -25,7 +25,7 @@ class Moderation < OpenAI::BaseModel # Whether any of the below categories are flagged. # # @return [Boolean] - required :flagged, OpenAI::BooleanModel + required :flagged, OpenAI::Internal::Type::BooleanModel # @!parse # # @param categories [OpenAI::Models::Moderation::Categories] @@ -35,23 +35,25 @@ class Moderation < OpenAI::BaseModel # # # def initialize(categories:, category_applied_input_types:, category_scores:, flagged:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Moderation#categories - class Categories < OpenAI::BaseModel + class Categories < OpenAI::Internal::Type::BaseModel # @!attribute harassment # Content that expresses, incites, or promotes harassing language towards any # target. # # @return [Boolean] - required :harassment, OpenAI::BooleanModel + required :harassment, OpenAI::Internal::Type::BooleanModel # @!attribute harassment_threatening # Harassment content that also includes violence or serious harm towards any # target. # # @return [Boolean] - required :harassment_threatening, OpenAI::BooleanModel, api_name: :"harassment/threatening" + required :harassment_threatening, + OpenAI::Internal::Type::BooleanModel, + api_name: :"harassment/threatening" # @!attribute hate # Content that expresses, incites, or promotes hate based on race, gender, @@ -60,7 +62,7 @@ class Categories < OpenAI::BaseModel # harassment. # # @return [Boolean] - required :hate, OpenAI::BooleanModel + required :hate, OpenAI::Internal::Type::BooleanModel # @!attribute hate_threatening # Hateful content that also includes violence or serious harm towards the targeted @@ -68,7 +70,7 @@ class Categories < OpenAI::BaseModel # orientation, disability status, or caste. # # @return [Boolean] - required :hate_threatening, OpenAI::BooleanModel, api_name: :"hate/threatening" + required :hate_threatening, OpenAI::Internal::Type::BooleanModel, api_name: :"hate/threatening" # @!attribute illicit # Content that includes instructions or advice that facilitate the planning or @@ -76,7 +78,7 @@ class Categories < OpenAI::BaseModel # illicit acts. For example, "how to shoplift" would fit this category. # # @return [Boolean, nil] - required :illicit, OpenAI::BooleanModel, nil?: true + required :illicit, OpenAI::Internal::Type::BooleanModel, nil?: true # @!attribute illicit_violent # Content that includes instructions or advice that facilitate the planning or @@ -84,14 +86,17 @@ class Categories < OpenAI::BaseModel # instruction on the procurement of any weapon. # # @return [Boolean, nil] - required :illicit_violent, OpenAI::BooleanModel, api_name: :"illicit/violent", nil?: true + required :illicit_violent, + OpenAI::Internal::Type::BooleanModel, + api_name: :"illicit/violent", + nil?: true # @!attribute self_harm # Content that promotes, encourages, or depicts acts of self-harm, such as # suicide, cutting, and eating disorders. # # @return [Boolean] - required :self_harm, OpenAI::BooleanModel, api_name: :"self-harm" + required :self_harm, OpenAI::Internal::Type::BooleanModel, api_name: :"self-harm" # @!attribute self_harm_instructions # Content that encourages performing acts of self-harm, such as suicide, cutting, @@ -99,14 +104,16 @@ class Categories < OpenAI::BaseModel # acts. # # @return [Boolean] - required :self_harm_instructions, OpenAI::BooleanModel, api_name: :"self-harm/instructions" + required :self_harm_instructions, + OpenAI::Internal::Type::BooleanModel, + api_name: :"self-harm/instructions" # @!attribute self_harm_intent # Content where the speaker expresses that they are engaging or intend to engage # in acts of self-harm, such as suicide, cutting, and eating disorders. # # @return [Boolean] - required :self_harm_intent, OpenAI::BooleanModel, api_name: :"self-harm/intent" + required :self_harm_intent, OpenAI::Internal::Type::BooleanModel, api_name: :"self-harm/intent" # @!attribute sexual # Content meant to arouse sexual excitement, such as the description of sexual @@ -114,25 +121,25 @@ class Categories < OpenAI::BaseModel # wellness). # # @return [Boolean] - required :sexual, OpenAI::BooleanModel + required :sexual, OpenAI::Internal::Type::BooleanModel # @!attribute sexual_minors # Sexual content that includes an individual who is under 18 years old. # # @return [Boolean] - required :sexual_minors, OpenAI::BooleanModel, api_name: :"sexual/minors" + required :sexual_minors, OpenAI::Internal::Type::BooleanModel, api_name: :"sexual/minors" # @!attribute violence # Content that depicts death, violence, or physical injury. # # @return [Boolean] - required :violence, OpenAI::BooleanModel + required :violence, OpenAI::Internal::Type::BooleanModel # @!attribute violence_graphic # Content that depicts death, violence, or physical injury in graphic detail. # # @return [Boolean] - required :violence_graphic, OpenAI::BooleanModel, api_name: :"violence/graphic" + required :violence_graphic, OpenAI::Internal::Type::BooleanModel, api_name: :"violence/graphic" # @!parse # # A list of the categories, and whether they are flagged or not. @@ -170,38 +177,39 @@ class Categories < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @see OpenAI::Models::Moderation#category_applied_input_types - class CategoryAppliedInputTypes < OpenAI::BaseModel + class CategoryAppliedInputTypes < OpenAI::Internal::Type::BaseModel # @!attribute harassment # The applied input type(s) for the category 'harassment'. # # @return [Array] required :harassment, - -> { OpenAI::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::Harassment] } + -> { OpenAI::Internal::Type::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::Harassment] } # @!attribute harassment_threatening # The applied input type(s) for the category 'harassment/threatening'. # # @return [Array] required :harassment_threatening, - -> { OpenAI::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::HarassmentThreatening] }, + -> { OpenAI::Internal::Type::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::HarassmentThreatening] }, api_name: :"harassment/threatening" # @!attribute hate # The applied input type(s) for the category 'hate'. # # @return [Array] - required :hate, -> { OpenAI::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::Hate] } + required :hate, + -> { OpenAI::Internal::Type::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::Hate] } # @!attribute hate_threatening # The applied input type(s) for the category 'hate/threatening'. # # @return [Array] required :hate_threatening, - -> { OpenAI::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::HateThreatening] }, + -> { OpenAI::Internal::Type::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::HateThreatening] }, api_name: :"hate/threatening" # @!attribute illicit @@ -209,14 +217,14 @@ class CategoryAppliedInputTypes < OpenAI::BaseModel # # @return [Array] required :illicit, - -> { OpenAI::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::Illicit] } + -> { OpenAI::Internal::Type::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::Illicit] } # @!attribute illicit_violent # The applied input type(s) for the category 'illicit/violent'. # # @return [Array] required :illicit_violent, - -> { OpenAI::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::IllicitViolent] }, + -> { OpenAI::Internal::Type::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::IllicitViolent] }, api_name: :"illicit/violent" # @!attribute self_harm @@ -224,7 +232,7 @@ class CategoryAppliedInputTypes < OpenAI::BaseModel # # @return [Array] required :self_harm, - -> { OpenAI::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::SelfHarm] }, + -> { OpenAI::Internal::Type::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::SelfHarm] }, api_name: :"self-harm" # @!attribute self_harm_instructions @@ -232,7 +240,7 @@ class CategoryAppliedInputTypes < OpenAI::BaseModel # # @return [Array] required :self_harm_instructions, - -> { OpenAI::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::SelfHarmInstruction] }, + -> { OpenAI::Internal::Type::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::SelfHarmInstruction] }, api_name: :"self-harm/instructions" # @!attribute self_harm_intent @@ -240,7 +248,7 @@ class CategoryAppliedInputTypes < OpenAI::BaseModel # # @return [Array] required :self_harm_intent, - -> { OpenAI::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::SelfHarmIntent] }, + -> { OpenAI::Internal::Type::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::SelfHarmIntent] }, api_name: :"self-harm/intent" # @!attribute sexual @@ -248,14 +256,14 @@ class CategoryAppliedInputTypes < OpenAI::BaseModel # # @return [Array] required :sexual, - -> { OpenAI::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::Sexual] } + -> { OpenAI::Internal::Type::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::Sexual] } # @!attribute sexual_minors # The applied input type(s) for the category 'sexual/minors'. # # @return [Array] required :sexual_minors, - -> { OpenAI::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::SexualMinor] }, + -> { OpenAI::Internal::Type::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::SexualMinor] }, api_name: :"sexual/minors" # @!attribute violence @@ -263,14 +271,14 @@ class CategoryAppliedInputTypes < OpenAI::BaseModel # # @return [Array] required :violence, - -> { OpenAI::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::Violence] } + -> { OpenAI::Internal::Type::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::Violence] } # @!attribute violence_graphic # The applied input type(s) for the category 'violence/graphic'. # # @return [Array] required :violence_graphic, - -> { OpenAI::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::ViolenceGraphic] }, + -> { OpenAI::Internal::Type::ArrayOf[enum: OpenAI::Models::Moderation::CategoryAppliedInputTypes::ViolenceGraphic] }, api_name: :"violence/graphic" # @!parse @@ -309,10 +317,10 @@ class CategoryAppliedInputTypes < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void module Harassment - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT = :text @@ -324,7 +332,7 @@ module Harassment end module HarassmentThreatening - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT = :text @@ -336,7 +344,7 @@ module HarassmentThreatening end module Hate - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT = :text @@ -348,7 +356,7 @@ module Hate end module HateThreatening - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT = :text @@ -360,7 +368,7 @@ module HateThreatening end module Illicit - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT = :text @@ -372,7 +380,7 @@ module Illicit end module IllicitViolent - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT = :text @@ -384,7 +392,7 @@ module IllicitViolent end module SelfHarm - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT = :text IMAGE = :image @@ -397,7 +405,7 @@ module SelfHarm end module SelfHarmInstruction - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT = :text IMAGE = :image @@ -410,7 +418,7 @@ module SelfHarmInstruction end module SelfHarmIntent - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT = :text IMAGE = :image @@ -423,7 +431,7 @@ module SelfHarmIntent end module Sexual - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT = :text IMAGE = :image @@ -436,7 +444,7 @@ module Sexual end module SexualMinor - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT = :text @@ -448,7 +456,7 @@ module SexualMinor end module Violence - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT = :text IMAGE = :image @@ -461,7 +469,7 @@ module Violence end module ViolenceGraphic - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT = :text IMAGE = :image @@ -475,7 +483,7 @@ module ViolenceGraphic end # @see OpenAI::Models::Moderation#category_scores - class CategoryScores < OpenAI::BaseModel + class CategoryScores < OpenAI::Internal::Type::BaseModel # @!attribute harassment # The score for the category 'harassment'. # @@ -590,7 +598,7 @@ class CategoryScores < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/moderation_create_params.rb b/lib/openai/models/moderation_create_params.rb index 95492506..c5e3ff4a 100644 --- a/lib/openai/models/moderation_create_params.rb +++ b/lib/openai/models/moderation_create_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Moderations#create - class ModerationCreateParams < OpenAI::BaseModel + class ModerationCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -35,12 +35,12 @@ class ModerationCreateParams < OpenAI::BaseModel # # # def initialize(input:, model: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Input (or inputs) to classify. Can be a single string, an array of strings, or # an array of multi-modal input objects similar to other models. module Input - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # A string of text to classify for moderation. variant String @@ -55,9 +55,10 @@ module Input # # @return [Array(String, Array, Array)] # def self.variants; end - StringArray = OpenAI::ArrayOf[String] + StringArray = OpenAI::Internal::Type::ArrayOf[String] - ModerationMultiModalInputArray = OpenAI::ArrayOf[union: -> { OpenAI::Models::ModerationMultiModalInput }] + ModerationMultiModalInputArray = + OpenAI::Internal::Type::ArrayOf[union: -> { OpenAI::Models::ModerationMultiModalInput }] end # The content moderation model you would like to use. Learn more in @@ -65,7 +66,7 @@ module Input # learn about available models # [here](https://platform.openai.com/docs/models#moderation). module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String diff --git a/lib/openai/models/moderation_create_response.rb b/lib/openai/models/moderation_create_response.rb index c06f6c57..922bca1e 100644 --- a/lib/openai/models/moderation_create_response.rb +++ b/lib/openai/models/moderation_create_response.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Moderations#create - class ModerationCreateResponse < OpenAI::BaseModel + class ModerationCreateResponse < OpenAI::Internal::Type::BaseModel # @!attribute id # The unique identifier for the moderation request. # @@ -20,7 +20,7 @@ class ModerationCreateResponse < OpenAI::BaseModel # A list of moderation objects. # # @return [Array] - required :results, -> { OpenAI::ArrayOf[OpenAI::Models::Moderation] } + required :results, -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Moderation] } # @!parse # # Represents if a given text input is potentially harmful. @@ -31,7 +31,7 @@ class ModerationCreateResponse < OpenAI::BaseModel # # # def initialize(id:, model:, results:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/moderation_image_url_input.rb b/lib/openai/models/moderation_image_url_input.rb index c02f0627..25b0835a 100644 --- a/lib/openai/models/moderation_image_url_input.rb +++ b/lib/openai/models/moderation_image_url_input.rb @@ -2,7 +2,7 @@ module OpenAI module Models - class ModerationImageURLInput < OpenAI::BaseModel + class ModerationImageURLInput < OpenAI::Internal::Type::BaseModel # @!attribute image_url # Contains either an image URL or a data URL for a base64 encoded image. # @@ -23,10 +23,10 @@ class ModerationImageURLInput < OpenAI::BaseModel # # # def initialize(image_url:, type: :image_url, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::ModerationImageURLInput#image_url - class ImageURL < OpenAI::BaseModel + class ImageURL < OpenAI::Internal::Type::BaseModel # @!attribute url # Either a URL of the image or the base64 encoded image data. # @@ -40,7 +40,7 @@ class ImageURL < OpenAI::BaseModel # # # def initialize(url:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/moderation_model.rb b/lib/openai/models/moderation_model.rb index 8d50271d..fa606daa 100644 --- a/lib/openai/models/moderation_model.rb +++ b/lib/openai/models/moderation_model.rb @@ -3,7 +3,7 @@ module OpenAI module Models module ModerationModel - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum OMNI_MODERATION_LATEST = :"omni-moderation-latest" OMNI_MODERATION_2024_09_26 = :"omni-moderation-2024-09-26" diff --git a/lib/openai/models/moderation_multi_modal_input.rb b/lib/openai/models/moderation_multi_modal_input.rb index 732739a6..ca2979e7 100644 --- a/lib/openai/models/moderation_multi_modal_input.rb +++ b/lib/openai/models/moderation_multi_modal_input.rb @@ -4,7 +4,7 @@ module OpenAI module Models # An object describing an image to classify. module ModerationMultiModalInput - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type diff --git a/lib/openai/models/moderation_text_input.rb b/lib/openai/models/moderation_text_input.rb index fbc192e8..ec5fe7cb 100644 --- a/lib/openai/models/moderation_text_input.rb +++ b/lib/openai/models/moderation_text_input.rb @@ -2,7 +2,7 @@ module OpenAI module Models - class ModerationTextInput < OpenAI::BaseModel + class ModerationTextInput < OpenAI::Internal::Type::BaseModel # @!attribute text # A string of text to classify. # @@ -23,7 +23,7 @@ class ModerationTextInput < OpenAI::BaseModel # # # def initialize(text:, type: :text, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/other_file_chunking_strategy_object.rb b/lib/openai/models/other_file_chunking_strategy_object.rb index 0f7c6224..9b28e285 100644 --- a/lib/openai/models/other_file_chunking_strategy_object.rb +++ b/lib/openai/models/other_file_chunking_strategy_object.rb @@ -2,7 +2,7 @@ module OpenAI module Models - class OtherFileChunkingStrategyObject < OpenAI::BaseModel + class OtherFileChunkingStrategyObject < OpenAI::Internal::Type::BaseModel # @!attribute type # Always `other`. # @@ -18,7 +18,7 @@ class OtherFileChunkingStrategyObject < OpenAI::BaseModel # # # def initialize(type: :other, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/reasoning.rb b/lib/openai/models/reasoning.rb index af30c4e0..0727d5c1 100644 --- a/lib/openai/models/reasoning.rb +++ b/lib/openai/models/reasoning.rb @@ -2,7 +2,7 @@ module OpenAI module Models - class Reasoning < OpenAI::BaseModel + class Reasoning < OpenAI::Internal::Type::BaseModel # @!attribute effort # **o-series models only** # @@ -35,7 +35,7 @@ class Reasoning < OpenAI::BaseModel # # # def initialize(effort: nil, generate_summary: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # **computer_use_preview only** # @@ -45,7 +45,7 @@ class Reasoning < OpenAI::BaseModel # # @see OpenAI::Models::Reasoning#generate_summary module GenerateSummary - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum CONCISE = :concise DETAILED = :detailed diff --git a/lib/openai/models/reasoning_effort.rb b/lib/openai/models/reasoning_effort.rb index f123316a..737f9d5f 100644 --- a/lib/openai/models/reasoning_effort.rb +++ b/lib/openai/models/reasoning_effort.rb @@ -9,7 +9,7 @@ module Models # supported values are `low`, `medium`, and `high`. Reducing reasoning effort can # result in faster responses and fewer tokens used on reasoning in a response. module ReasoningEffort - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum LOW = :low MEDIUM = :medium diff --git a/lib/openai/models/response_format_json_object.rb b/lib/openai/models/response_format_json_object.rb index 2996332d..bcd20439 100644 --- a/lib/openai/models/response_format_json_object.rb +++ b/lib/openai/models/response_format_json_object.rb @@ -2,7 +2,7 @@ module OpenAI module Models - class ResponseFormatJSONObject < OpenAI::BaseModel + class ResponseFormatJSONObject < OpenAI::Internal::Type::BaseModel # @!attribute type # The type of response format being defined. Always `json_object`. # @@ -18,7 +18,7 @@ class ResponseFormatJSONObject < OpenAI::BaseModel # # # def initialize(type: :json_object, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/response_format_json_schema.rb b/lib/openai/models/response_format_json_schema.rb index 8dac86a9..71050de4 100644 --- a/lib/openai/models/response_format_json_schema.rb +++ b/lib/openai/models/response_format_json_schema.rb @@ -2,7 +2,7 @@ module OpenAI module Models - class ResponseFormatJSONSchema < OpenAI::BaseModel + class ResponseFormatJSONSchema < OpenAI::Internal::Type::BaseModel # @!attribute json_schema # Structured Outputs configuration options, including a JSON Schema. # @@ -25,10 +25,10 @@ class ResponseFormatJSONSchema < OpenAI::BaseModel # # # def initialize(json_schema:, type: :json_schema, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::ResponseFormatJSONSchema#json_schema - class JSONSchema < OpenAI::BaseModel + class JSONSchema < OpenAI::Internal::Type::BaseModel # @!attribute name # The name of the response format. Must be a-z, A-Z, 0-9, or contain underscores # and dashes, with a maximum length of 64. @@ -52,7 +52,7 @@ class JSONSchema < OpenAI::BaseModel # to build JSON schemas [here](https://json-schema.org/). # # @return [Hash{Symbol=>Object}, nil] - optional :schema, OpenAI::HashOf[OpenAI::Unknown] + optional :schema, OpenAI::Internal::Type::HashOf[OpenAI::Internal::Type::Unknown] # @!parse # # @return [Hash{Symbol=>Object}] @@ -66,7 +66,7 @@ class JSONSchema < OpenAI::BaseModel # [Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs). # # @return [Boolean, nil] - optional :strict, OpenAI::BooleanModel, nil?: true + optional :strict, OpenAI::Internal::Type::BooleanModel, nil?: true # @!parse # # Structured Outputs configuration options, including a JSON Schema. @@ -78,7 +78,7 @@ class JSONSchema < OpenAI::BaseModel # # # def initialize(name:, description: nil, schema: nil, strict: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/response_format_text.rb b/lib/openai/models/response_format_text.rb index 3821c9a0..609679d6 100644 --- a/lib/openai/models/response_format_text.rb +++ b/lib/openai/models/response_format_text.rb @@ -2,7 +2,7 @@ module OpenAI module Models - class ResponseFormatText < OpenAI::BaseModel + class ResponseFormatText < OpenAI::Internal::Type::BaseModel # @!attribute type # The type of response format being defined. Always `text`. # @@ -16,7 +16,7 @@ class ResponseFormatText < OpenAI::BaseModel # # # def initialize(type: :text, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/computer_tool.rb b/lib/openai/models/responses/computer_tool.rb index 57d62fd4..dcbbba28 100644 --- a/lib/openai/models/responses/computer_tool.rb +++ b/lib/openai/models/responses/computer_tool.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ComputerTool < OpenAI::BaseModel + class ComputerTool < OpenAI::Internal::Type::BaseModel # @!attribute display_height # The height of the computer display. # @@ -39,13 +39,13 @@ class ComputerTool < OpenAI::BaseModel # # # def initialize(display_height:, display_width:, environment:, type: :computer_use_preview, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The type of computer environment to control. # # @see OpenAI::Models::Responses::ComputerTool#environment module Environment - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum MAC = :mac WINDOWS = :windows diff --git a/lib/openai/models/responses/easy_input_message.rb b/lib/openai/models/responses/easy_input_message.rb index 715855e0..d7c78f66 100644 --- a/lib/openai/models/responses/easy_input_message.rb +++ b/lib/openai/models/responses/easy_input_message.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class EasyInputMessage < OpenAI::BaseModel + class EasyInputMessage < OpenAI::Internal::Type::BaseModel # @!attribute content # Text, image, or audio input to the model, used to generate a response. Can also # contain previous assistant responses. @@ -41,14 +41,14 @@ class EasyInputMessage < OpenAI::BaseModel # # # def initialize(content:, role:, type: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Text, image, or audio input to the model, used to generate a response. Can also # contain previous assistant responses. # # @see OpenAI::Models::Responses::EasyInputMessage#content module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # A text input to the model. variant String @@ -67,7 +67,7 @@ module Content # # @see OpenAI::Models::Responses::EasyInputMessage#role module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum USER = :user ASSISTANT = :assistant @@ -85,7 +85,7 @@ module Role # # @see OpenAI::Models::Responses::EasyInputMessage#type module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum MESSAGE = :message diff --git a/lib/openai/models/responses/file_search_tool.rb b/lib/openai/models/responses/file_search_tool.rb index 18833fef..62d497ad 100644 --- a/lib/openai/models/responses/file_search_tool.rb +++ b/lib/openai/models/responses/file_search_tool.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class FileSearchTool < OpenAI::BaseModel + class FileSearchTool < OpenAI::Internal::Type::BaseModel # @!attribute type # The type of the file search tool. Always `file_search`. # @@ -14,7 +14,7 @@ class FileSearchTool < OpenAI::BaseModel # The IDs of the vector stores to search. # # @return [Array] - required :vector_store_ids, OpenAI::ArrayOf[String] + required :vector_store_ids, OpenAI::Internal::Type::ArrayOf[String] # @!attribute [r] filters # A filter to apply based on file attributes. @@ -60,13 +60,13 @@ class FileSearchTool < OpenAI::BaseModel # # # def initialize(vector_store_ids:, filters: nil, max_num_results: nil, ranking_options: nil, type: :file_search, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # A filter to apply based on file attributes. # # @see OpenAI::Models::Responses::FileSearchTool#filters module Filters - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # A filter used to compare a specified attribute key to a given value using a defined comparison operation. variant -> { OpenAI::Models::ComparisonFilter } @@ -80,7 +80,7 @@ module Filters end # @see OpenAI::Models::Responses::FileSearchTool#ranking_options - class RankingOptions < OpenAI::BaseModel + class RankingOptions < OpenAI::Internal::Type::BaseModel # @!attribute [r] ranker # The ranker to use for the file search. # @@ -111,13 +111,13 @@ class RankingOptions < OpenAI::BaseModel # # # def initialize(ranker: nil, score_threshold: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The ranker to use for the file search. # # @see OpenAI::Models::Responses::FileSearchTool::RankingOptions#ranker module Ranker - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO = :auto DEFAULT_2024_11_15 = :"default-2024-11-15" diff --git a/lib/openai/models/responses/function_tool.rb b/lib/openai/models/responses/function_tool.rb index 5e1906c6..ce5b34a5 100644 --- a/lib/openai/models/responses/function_tool.rb +++ b/lib/openai/models/responses/function_tool.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class FunctionTool < OpenAI::BaseModel + class FunctionTool < OpenAI::Internal::Type::BaseModel # @!attribute name # The name of the function to call. # @@ -14,13 +14,13 @@ class FunctionTool < OpenAI::BaseModel # A JSON schema object describing the parameters of the function. # # @return [Hash{Symbol=>Object}] - required :parameters, OpenAI::HashOf[OpenAI::Unknown] + required :parameters, OpenAI::Internal::Type::HashOf[OpenAI::Internal::Type::Unknown] # @!attribute strict # Whether to enforce strict parameter validation. Default `true`. # # @return [Boolean] - required :strict, OpenAI::BooleanModel + required :strict, OpenAI::Internal::Type::BooleanModel # @!attribute type # The type of the function tool. Always `function`. @@ -48,7 +48,7 @@ class FunctionTool < OpenAI::BaseModel # # # def initialize(name:, parameters:, strict:, description: nil, type: :function, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/input_item_list_params.rb b/lib/openai/models/responses/input_item_list_params.rb index b807321f..a6a7a77f 100644 --- a/lib/openai/models/responses/input_item_list_params.rb +++ b/lib/openai/models/responses/input_item_list_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Responses # @see OpenAI::Resources::Responses::InputItems#list - class InputItemListParams < OpenAI::BaseModel + class InputItemListParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -34,7 +34,8 @@ class InputItemListParams < OpenAI::BaseModel # Response creation above for more information. # # @return [Array, nil] - optional :include, -> { OpenAI::ArrayOf[enum: OpenAI::Models::Responses::ResponseIncludable] } + optional :include, + -> { OpenAI::Internal::Type::ArrayOf[enum: OpenAI::Models::Responses::ResponseIncludable] } # @!parse # # @return [Array] @@ -74,14 +75,14 @@ class InputItemListParams < OpenAI::BaseModel # # # def initialize(after: nil, before: nil, include: nil, limit: nil, order: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The order to return the input items in. Default is `asc`. # # - `asc`: Return the input items in ascending order. # - `desc`: Return the input items in descending order. module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASC = :asc DESC = :desc diff --git a/lib/openai/models/responses/response.rb b/lib/openai/models/responses/response.rb index b6e167a0..afcc7eec 100644 --- a/lib/openai/models/responses/response.rb +++ b/lib/openai/models/responses/response.rb @@ -6,7 +6,7 @@ module Responses # @see OpenAI::Resources::Responses#create # # @see OpenAI::Resources::Responses#stream_raw - class Response < OpenAI::BaseModel + class Response < OpenAI::Internal::Type::BaseModel # @!attribute id # Unique identifier for this Response. # @@ -51,7 +51,7 @@ class Response < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - required :metadata, OpenAI::HashOf[String], nil?: true + required :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!attribute model # Model ID used to generate the response, like `gpt-4o` or `o1`. OpenAI offers a @@ -79,13 +79,14 @@ class Response < OpenAI::BaseModel # consider using the `output_text` property where supported in SDKs. # # @return [Array] - required :output, -> { OpenAI::ArrayOf[union: OpenAI::Models::Responses::ResponseOutputItem] } + required :output, + -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Responses::ResponseOutputItem] } # @!attribute parallel_tool_calls # Whether to allow the model to run tool calls in parallel. # # @return [Boolean] - required :parallel_tool_calls, OpenAI::BooleanModel + required :parallel_tool_calls, OpenAI::Internal::Type::BooleanModel # @!attribute temperature # What sampling temperature to use, between 0 and 2. Higher values like 0.8 will @@ -121,7 +122,7 @@ class Response < OpenAI::BaseModel # [function calling](https://platform.openai.com/docs/guides/function-calling). # # @return [Array] - required :tools, -> { OpenAI::ArrayOf[union: OpenAI::Models::Responses::Tool] } + required :tools, -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Responses::Tool] } # @!attribute top_p # An alternative to sampling with temperature, called nucleus sampling, where the @@ -270,10 +271,10 @@ class Response < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Responses::Response#incomplete_details - class IncompleteDetails < OpenAI::BaseModel + class IncompleteDetails < OpenAI::Internal::Type::BaseModel # @!attribute [r] reason # The reason why the response is incomplete. # @@ -291,13 +292,13 @@ class IncompleteDetails < OpenAI::BaseModel # # # def initialize(reason: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The reason why the response is incomplete. # # @see OpenAI::Models::Responses::Response::IncompleteDetails#reason module Reason - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum MAX_OUTPUT_TOKENS = :max_output_tokens CONTENT_FILTER = :content_filter @@ -316,7 +317,7 @@ module Reason # # @see OpenAI::Models::Responses::Response#tool_choice module ToolChoice - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # Controls which (if any) tool is called by the model. # @@ -350,7 +351,7 @@ module ToolChoice # # @see OpenAI::Models::Responses::Response#truncation module Truncation - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO = :auto DISABLED = :disabled diff --git a/lib/openai/models/responses/response_audio_delta_event.rb b/lib/openai/models/responses/response_audio_delta_event.rb index 8c690aef..72fd5781 100644 --- a/lib/openai/models/responses/response_audio_delta_event.rb +++ b/lib/openai/models/responses/response_audio_delta_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseAudioDeltaEvent < OpenAI::BaseModel + class ResponseAudioDeltaEvent < OpenAI::Internal::Type::BaseModel # @!attribute delta # A chunk of Base64 encoded response audio bytes. # @@ -24,7 +24,7 @@ class ResponseAudioDeltaEvent < OpenAI::BaseModel # # # def initialize(delta:, type: :"response.audio.delta", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_audio_done_event.rb b/lib/openai/models/responses/response_audio_done_event.rb index e0f3632f..6dd92e68 100644 --- a/lib/openai/models/responses/response_audio_done_event.rb +++ b/lib/openai/models/responses/response_audio_done_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseAudioDoneEvent < OpenAI::BaseModel + class ResponseAudioDoneEvent < OpenAI::Internal::Type::BaseModel # @!attribute type # The type of the event. Always `response.audio.done`. # @@ -17,7 +17,7 @@ class ResponseAudioDoneEvent < OpenAI::BaseModel # # # def initialize(type: :"response.audio.done", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_audio_transcript_delta_event.rb b/lib/openai/models/responses/response_audio_transcript_delta_event.rb index 476cb265..062c5a76 100644 --- a/lib/openai/models/responses/response_audio_transcript_delta_event.rb +++ b/lib/openai/models/responses/response_audio_transcript_delta_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseAudioTranscriptDeltaEvent < OpenAI::BaseModel + class ResponseAudioTranscriptDeltaEvent < OpenAI::Internal::Type::BaseModel # @!attribute delta # The partial transcript of the audio response. # @@ -24,7 +24,7 @@ class ResponseAudioTranscriptDeltaEvent < OpenAI::BaseModel # # # def initialize(delta:, type: :"response.audio.transcript.delta", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_audio_transcript_done_event.rb b/lib/openai/models/responses/response_audio_transcript_done_event.rb index 89ee6d86..98399c6e 100644 --- a/lib/openai/models/responses/response_audio_transcript_done_event.rb +++ b/lib/openai/models/responses/response_audio_transcript_done_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseAudioTranscriptDoneEvent < OpenAI::BaseModel + class ResponseAudioTranscriptDoneEvent < OpenAI::Internal::Type::BaseModel # @!attribute type # The type of the event. Always `response.audio.transcript.done`. # @@ -17,7 +17,7 @@ class ResponseAudioTranscriptDoneEvent < OpenAI::BaseModel # # # def initialize(type: :"response.audio.transcript.done", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_code_interpreter_call_code_delta_event.rb b/lib/openai/models/responses/response_code_interpreter_call_code_delta_event.rb index 3ec0bcba..2fa390fe 100644 --- a/lib/openai/models/responses/response_code_interpreter_call_code_delta_event.rb +++ b/lib/openai/models/responses/response_code_interpreter_call_code_delta_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseCodeInterpreterCallCodeDeltaEvent < OpenAI::BaseModel + class ResponseCodeInterpreterCallCodeDeltaEvent < OpenAI::Internal::Type::BaseModel # @!attribute delta # The partial code snippet added by the code interpreter. # @@ -31,7 +31,7 @@ class ResponseCodeInterpreterCallCodeDeltaEvent < OpenAI::BaseModel # # # def initialize(delta:, output_index:, type: :"response.code_interpreter_call.code.delta", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_code_interpreter_call_code_done_event.rb b/lib/openai/models/responses/response_code_interpreter_call_code_done_event.rb index 19de0973..caa291b9 100644 --- a/lib/openai/models/responses/response_code_interpreter_call_code_done_event.rb +++ b/lib/openai/models/responses/response_code_interpreter_call_code_done_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseCodeInterpreterCallCodeDoneEvent < OpenAI::BaseModel + class ResponseCodeInterpreterCallCodeDoneEvent < OpenAI::Internal::Type::BaseModel # @!attribute code # The final code snippet output by the code interpreter. # @@ -31,7 +31,7 @@ class ResponseCodeInterpreterCallCodeDoneEvent < OpenAI::BaseModel # # # def initialize(code:, output_index:, type: :"response.code_interpreter_call.code.done", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_code_interpreter_call_completed_event.rb b/lib/openai/models/responses/response_code_interpreter_call_completed_event.rb index 3843f408..0e13a5e9 100644 --- a/lib/openai/models/responses/response_code_interpreter_call_completed_event.rb +++ b/lib/openai/models/responses/response_code_interpreter_call_completed_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseCodeInterpreterCallCompletedEvent < OpenAI::BaseModel + class ResponseCodeInterpreterCallCompletedEvent < OpenAI::Internal::Type::BaseModel # @!attribute code_interpreter_call # A tool call to run code. # @@ -31,7 +31,7 @@ class ResponseCodeInterpreterCallCompletedEvent < OpenAI::BaseModel # # # def initialize(code_interpreter_call:, output_index:, type: :"response.code_interpreter_call.completed", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_code_interpreter_call_in_progress_event.rb b/lib/openai/models/responses/response_code_interpreter_call_in_progress_event.rb index 549d7eba..66aa4fce 100644 --- a/lib/openai/models/responses/response_code_interpreter_call_in_progress_event.rb +++ b/lib/openai/models/responses/response_code_interpreter_call_in_progress_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseCodeInterpreterCallInProgressEvent < OpenAI::BaseModel + class ResponseCodeInterpreterCallInProgressEvent < OpenAI::Internal::Type::BaseModel # @!attribute code_interpreter_call # A tool call to run code. # @@ -31,7 +31,7 @@ class ResponseCodeInterpreterCallInProgressEvent < OpenAI::BaseModel # # # def initialize(code_interpreter_call:, output_index:, type: :"response.code_interpreter_call.in_progress", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_code_interpreter_call_interpreting_event.rb b/lib/openai/models/responses/response_code_interpreter_call_interpreting_event.rb index bf8bbe13..7a87f2c2 100644 --- a/lib/openai/models/responses/response_code_interpreter_call_interpreting_event.rb +++ b/lib/openai/models/responses/response_code_interpreter_call_interpreting_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseCodeInterpreterCallInterpretingEvent < OpenAI::BaseModel + class ResponseCodeInterpreterCallInterpretingEvent < OpenAI::Internal::Type::BaseModel # @!attribute code_interpreter_call # A tool call to run code. # @@ -31,7 +31,7 @@ class ResponseCodeInterpreterCallInterpretingEvent < OpenAI::BaseModel # # # def initialize(code_interpreter_call:, output_index:, type: :"response.code_interpreter_call.interpreting", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_code_interpreter_tool_call.rb b/lib/openai/models/responses/response_code_interpreter_tool_call.rb index 5ce7e550..9d7260d9 100644 --- a/lib/openai/models/responses/response_code_interpreter_tool_call.rb +++ b/lib/openai/models/responses/response_code_interpreter_tool_call.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseCodeInterpreterToolCall < OpenAI::BaseModel + class ResponseCodeInterpreterToolCall < OpenAI::Internal::Type::BaseModel # @!attribute id # The unique ID of the code interpreter tool call. # @@ -21,7 +21,7 @@ class ResponseCodeInterpreterToolCall < OpenAI::BaseModel # # @return [Array] required :results, - -> { OpenAI::ArrayOf[union: OpenAI::Models::Responses::ResponseCodeInterpreterToolCall::Result] } + -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Responses::ResponseCodeInterpreterToolCall::Result] } # @!attribute status # The status of the code interpreter tool call. @@ -46,11 +46,11 @@ class ResponseCodeInterpreterToolCall < OpenAI::BaseModel # # # def initialize(id:, code:, results:, status:, type: :code_interpreter_call, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The output of a code interpreter tool call that is text. module Result - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type @@ -60,7 +60,7 @@ module Result # The output of a code interpreter tool call that is a file. variant :files, -> { OpenAI::Models::Responses::ResponseCodeInterpreterToolCall::Result::Files } - class Logs < OpenAI::BaseModel + class Logs < OpenAI::Internal::Type::BaseModel # @!attribute logs # The logs of the code interpreter tool call. # @@ -81,15 +81,15 @@ class Logs < OpenAI::BaseModel # # # def initialize(logs:, type: :logs, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class Files < OpenAI::BaseModel + class Files < OpenAI::Internal::Type::BaseModel # @!attribute files # # @return [Array] required :files, - -> { OpenAI::ArrayOf[OpenAI::Models::Responses::ResponseCodeInterpreterToolCall::Result::Files::File] } + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Responses::ResponseCodeInterpreterToolCall::Result::Files::File] } # @!attribute type # The type of the code interpreter file output. Always `files`. @@ -105,9 +105,9 @@ class Files < OpenAI::BaseModel # # # def initialize(files:, type: :files, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class File < OpenAI::BaseModel + class File < OpenAI::Internal::Type::BaseModel # @!attribute file_id # The ID of the file. # @@ -126,7 +126,7 @@ class File < OpenAI::BaseModel # # # def initialize(file_id:, mime_type:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end @@ -139,7 +139,7 @@ class File < OpenAI::BaseModel # # @see OpenAI::Models::Responses::ResponseCodeInterpreterToolCall#status module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS = :in_progress INTERPRETING = :interpreting diff --git a/lib/openai/models/responses/response_completed_event.rb b/lib/openai/models/responses/response_completed_event.rb index 8328ff58..883a7de9 100644 --- a/lib/openai/models/responses/response_completed_event.rb +++ b/lib/openai/models/responses/response_completed_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseCompletedEvent < OpenAI::BaseModel + class ResponseCompletedEvent < OpenAI::Internal::Type::BaseModel # @!attribute response # Properties of the completed response. # @@ -24,7 +24,7 @@ class ResponseCompletedEvent < OpenAI::BaseModel # # # def initialize(response:, type: :"response.completed", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_computer_tool_call.rb b/lib/openai/models/responses/response_computer_tool_call.rb index 207a2584..7eafa499 100644 --- a/lib/openai/models/responses/response_computer_tool_call.rb +++ b/lib/openai/models/responses/response_computer_tool_call.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseComputerToolCall < OpenAI::BaseModel + class ResponseComputerToolCall < OpenAI::Internal::Type::BaseModel # @!attribute id # The unique ID of the computer call. # @@ -27,7 +27,7 @@ class ResponseComputerToolCall < OpenAI::BaseModel # # @return [Array] required :pending_safety_checks, - -> { OpenAI::ArrayOf[OpenAI::Models::Responses::ResponseComputerToolCall::PendingSafetyCheck] } + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Responses::ResponseComputerToolCall::PendingSafetyCheck] } # @!attribute status # The status of the item. One of `in_progress`, `completed`, or `incomplete`. @@ -56,13 +56,13 @@ class ResponseComputerToolCall < OpenAI::BaseModel # # # def initialize(id:, action:, call_id:, pending_safety_checks:, status:, type:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # A click action. # # @see OpenAI::Models::Responses::ResponseComputerToolCall#action module Action - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type @@ -93,7 +93,7 @@ module Action # A wait action. variant :wait, -> { OpenAI::Models::Responses::ResponseComputerToolCall::Action::Wait } - class Click < OpenAI::BaseModel + class Click < OpenAI::Internal::Type::BaseModel # @!attribute button # Indicates which mouse button was pressed during the click. One of `left`, # `right`, `wheel`, `back`, or `forward`. @@ -130,14 +130,14 @@ class Click < OpenAI::BaseModel # # # def initialize(button:, x:, y_:, type: :click, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Indicates which mouse button was pressed during the click. One of `left`, # `right`, `wheel`, `back`, or `forward`. # # @see OpenAI::Models::Responses::ResponseComputerToolCall::Action::Click#button module Button - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum LEFT = :left RIGHT = :right @@ -153,7 +153,7 @@ module Button end end - class DoubleClick < OpenAI::BaseModel + class DoubleClick < OpenAI::Internal::Type::BaseModel # @!attribute type # Specifies the event type. For a double click action, this property is always set # to `double_click`. @@ -182,10 +182,10 @@ class DoubleClick < OpenAI::BaseModel # # # def initialize(x:, y_:, type: :double_click, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class Drag < OpenAI::BaseModel + class Drag < OpenAI::Internal::Type::BaseModel # @!attribute path # An array of coordinates representing the path of the drag action. Coordinates # will appear as an array of objects, eg @@ -199,7 +199,7 @@ class Drag < OpenAI::BaseModel # # @return [Array] required :path, - -> { OpenAI::ArrayOf[OpenAI::Models::Responses::ResponseComputerToolCall::Action::Drag::Path] } + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Responses::ResponseComputerToolCall::Action::Drag::Path] } # @!attribute type # Specifies the event type. For a drag action, this property is always set to @@ -216,9 +216,9 @@ class Drag < OpenAI::BaseModel # # # def initialize(path:, type: :drag, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class Path < OpenAI::BaseModel + class Path < OpenAI::Internal::Type::BaseModel # @!attribute x # The x-coordinate. # @@ -239,17 +239,17 @@ class Path < OpenAI::BaseModel # # # def initialize(x:, y_:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end - class Keypress < OpenAI::BaseModel + class Keypress < OpenAI::Internal::Type::BaseModel # @!attribute keys # The combination of keys the model is requesting to be pressed. This is an array # of strings, each representing a key. # # @return [Array] - required :keys, OpenAI::ArrayOf[String] + required :keys, OpenAI::Internal::Type::ArrayOf[String] # @!attribute type # Specifies the event type. For a keypress action, this property is always set to @@ -266,10 +266,10 @@ class Keypress < OpenAI::BaseModel # # # def initialize(keys:, type: :keypress, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class Move < OpenAI::BaseModel + class Move < OpenAI::Internal::Type::BaseModel # @!attribute type # Specifies the event type. For a move action, this property is always set to # `move`. @@ -298,10 +298,10 @@ class Move < OpenAI::BaseModel # # # def initialize(x:, y_:, type: :move, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class Screenshot < OpenAI::BaseModel + class Screenshot < OpenAI::Internal::Type::BaseModel # @!attribute type # Specifies the event type. For a screenshot action, this property is always set # to `screenshot`. @@ -316,10 +316,10 @@ class Screenshot < OpenAI::BaseModel # # # def initialize(type: :screenshot, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class Scroll < OpenAI::BaseModel + class Scroll < OpenAI::Internal::Type::BaseModel # @!attribute scroll_x # The horizontal scroll distance. # @@ -362,10 +362,10 @@ class Scroll < OpenAI::BaseModel # # # def initialize(scroll_x:, scroll_y:, x:, y_:, type: :scroll, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class Type < OpenAI::BaseModel + class Type < OpenAI::Internal::Type::BaseModel # @!attribute text # The text to type. # @@ -387,10 +387,10 @@ class Type < OpenAI::BaseModel # # # def initialize(text:, type: :type, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class Wait < OpenAI::BaseModel + class Wait < OpenAI::Internal::Type::BaseModel # @!attribute type # Specifies the event type. For a wait action, this property is always set to # `wait`. @@ -405,7 +405,7 @@ class Wait < OpenAI::BaseModel # # # def initialize(type: :wait, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @!parse @@ -413,7 +413,7 @@ class Wait < OpenAI::BaseModel # def self.variants; end end - class PendingSafetyCheck < OpenAI::BaseModel + class PendingSafetyCheck < OpenAI::Internal::Type::BaseModel # @!attribute id # The ID of the pending safety check. # @@ -441,7 +441,7 @@ class PendingSafetyCheck < OpenAI::BaseModel # # # def initialize(id:, code:, message:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # The status of the item. One of `in_progress`, `completed`, or `incomplete`. @@ -449,7 +449,7 @@ class PendingSafetyCheck < OpenAI::BaseModel # # @see OpenAI::Models::Responses::ResponseComputerToolCall#status module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS = :in_progress COMPLETED = :completed @@ -466,7 +466,7 @@ module Status # # @see OpenAI::Models::Responses::ResponseComputerToolCall#type module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum COMPUTER_CALL = :computer_call diff --git a/lib/openai/models/responses/response_computer_tool_call_output_item.rb b/lib/openai/models/responses/response_computer_tool_call_output_item.rb index 467d41e0..0db1e5cc 100644 --- a/lib/openai/models/responses/response_computer_tool_call_output_item.rb +++ b/lib/openai/models/responses/response_computer_tool_call_output_item.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseComputerToolCallOutputItem < OpenAI::BaseModel + class ResponseComputerToolCallOutputItem < OpenAI::Internal::Type::BaseModel # @!attribute id # The unique ID of the computer call tool output. # @@ -34,7 +34,7 @@ class ResponseComputerToolCallOutputItem < OpenAI::BaseModel # # @return [Array, nil] optional :acknowledged_safety_checks, - -> { OpenAI::ArrayOf[OpenAI::Models::Responses::ResponseComputerToolCallOutputItem::AcknowledgedSafetyCheck] } + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Responses::ResponseComputerToolCallOutputItem::AcknowledgedSafetyCheck] } # @!parse # # @return [Array] @@ -61,9 +61,9 @@ class ResponseComputerToolCallOutputItem < OpenAI::BaseModel # # # def initialize(id:, call_id:, output:, acknowledged_safety_checks: nil, status: nil, type: :computer_call_output, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class AcknowledgedSafetyCheck < OpenAI::BaseModel + class AcknowledgedSafetyCheck < OpenAI::Internal::Type::BaseModel # @!attribute id # The ID of the pending safety check. # @@ -91,7 +91,7 @@ class AcknowledgedSafetyCheck < OpenAI::BaseModel # # # def initialize(id:, code:, message:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # The status of the message input. One of `in_progress`, `completed`, or @@ -99,7 +99,7 @@ class AcknowledgedSafetyCheck < OpenAI::BaseModel # # @see OpenAI::Models::Responses::ResponseComputerToolCallOutputItem#status module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS = :in_progress COMPLETED = :completed diff --git a/lib/openai/models/responses/response_computer_tool_call_output_screenshot.rb b/lib/openai/models/responses/response_computer_tool_call_output_screenshot.rb index 4dc426ff..1ac341fb 100644 --- a/lib/openai/models/responses/response_computer_tool_call_output_screenshot.rb +++ b/lib/openai/models/responses/response_computer_tool_call_output_screenshot.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseComputerToolCallOutputScreenshot < OpenAI::BaseModel + class ResponseComputerToolCallOutputScreenshot < OpenAI::Internal::Type::BaseModel # @!attribute type # Specifies the event type. For a computer screenshot, this property is always set # to `computer_screenshot`. @@ -40,7 +40,7 @@ class ResponseComputerToolCallOutputScreenshot < OpenAI::BaseModel # # # def initialize(file_id: nil, image_url: nil, type: :computer_screenshot, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_content.rb b/lib/openai/models/responses/response_content.rb index c3e86848..3c306ddb 100644 --- a/lib/openai/models/responses/response_content.rb +++ b/lib/openai/models/responses/response_content.rb @@ -5,7 +5,7 @@ module Models module Responses # Multi-modal input and output contents. module ResponseContent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # A text input to the model. variant -> { OpenAI::Models::Responses::ResponseInputText } diff --git a/lib/openai/models/responses/response_content_part_added_event.rb b/lib/openai/models/responses/response_content_part_added_event.rb index 66590848..2d78d1ff 100644 --- a/lib/openai/models/responses/response_content_part_added_event.rb +++ b/lib/openai/models/responses/response_content_part_added_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseContentPartAddedEvent < OpenAI::BaseModel + class ResponseContentPartAddedEvent < OpenAI::Internal::Type::BaseModel # @!attribute content_index # The index of the content part that was added. # @@ -45,13 +45,13 @@ class ResponseContentPartAddedEvent < OpenAI::BaseModel # # # def initialize(content_index:, item_id:, output_index:, part:, type: :"response.content_part.added", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The content part that was added. # # @see OpenAI::Models::Responses::ResponseContentPartAddedEvent#part module Part - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type diff --git a/lib/openai/models/responses/response_content_part_done_event.rb b/lib/openai/models/responses/response_content_part_done_event.rb index 988c11d1..d28334f9 100644 --- a/lib/openai/models/responses/response_content_part_done_event.rb +++ b/lib/openai/models/responses/response_content_part_done_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseContentPartDoneEvent < OpenAI::BaseModel + class ResponseContentPartDoneEvent < OpenAI::Internal::Type::BaseModel # @!attribute content_index # The index of the content part that is done. # @@ -45,13 +45,13 @@ class ResponseContentPartDoneEvent < OpenAI::BaseModel # # # def initialize(content_index:, item_id:, output_index:, part:, type: :"response.content_part.done", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The content part that is done. # # @see OpenAI::Models::Responses::ResponseContentPartDoneEvent#part module Part - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type diff --git a/lib/openai/models/responses/response_create_params.rb b/lib/openai/models/responses/response_create_params.rb index 8253c846..7015d62b 100644 --- a/lib/openai/models/responses/response_create_params.rb +++ b/lib/openai/models/responses/response_create_params.rb @@ -6,7 +6,7 @@ module Responses # @see OpenAI::Resources::Responses#create # # @see OpenAI::Resources::Responses#stream_raw - class ResponseCreateParams < OpenAI::BaseModel + class ResponseCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -46,7 +46,9 @@ class ResponseCreateParams < OpenAI::BaseModel # call output. # # @return [Array, nil] - optional :include, -> { OpenAI::ArrayOf[enum: OpenAI::Models::Responses::ResponseIncludable] }, nil?: true + optional :include, + -> { OpenAI::Internal::Type::ArrayOf[enum: OpenAI::Models::Responses::ResponseIncludable] }, + nil?: true # @!attribute instructions # Inserts a system (or developer) message as the first item in the model's @@ -76,13 +78,13 @@ class ResponseCreateParams < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!attribute parallel_tool_calls # Whether to allow the model to run tool calls in parallel. # # @return [Boolean, nil] - optional :parallel_tool_calls, OpenAI::BooleanModel, nil?: true + optional :parallel_tool_calls, OpenAI::Internal::Type::BooleanModel, nil?: true # @!attribute previous_response_id # The unique ID of the previous response to the model. Use this to create @@ -105,7 +107,7 @@ class ResponseCreateParams < OpenAI::BaseModel # Whether to store the generated model response for later retrieval via API. # # @return [Boolean, nil] - optional :store, OpenAI::BooleanModel, nil?: true + optional :store, OpenAI::Internal::Type::BooleanModel, nil?: true # @!attribute temperature # What sampling temperature to use, between 0 and 2. Higher values like 0.8 will @@ -159,7 +161,7 @@ class ResponseCreateParams < OpenAI::BaseModel # [function calling](https://platform.openai.com/docs/guides/function-calling). # # @return [Array, nil] - optional :tools, -> { OpenAI::ArrayOf[union: OpenAI::Models::Responses::Tool] } + optional :tools, -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Responses::Tool] } # @!parse # # @return [Array] @@ -243,7 +245,7 @@ class ResponseCreateParams < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Text, image, or file inputs to the model, used to generate a response. # @@ -255,7 +257,7 @@ class ResponseCreateParams < OpenAI::BaseModel # - [Conversation state](https://platform.openai.com/docs/guides/conversation-state) # - [Function calling](https://platform.openai.com/docs/guides/function-calling) module Input - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # A text input to the model, equivalent to a text input with the # `user` role. @@ -274,7 +276,7 @@ module Input # response. See the `tools` parameter to see how to specify which tools the model # can call. module ToolChoice - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # Controls which (if any) tool is called by the model. # @@ -306,7 +308,7 @@ module ToolChoice # - `disabled` (default): If a model response will exceed the context window size # for a model, the request will fail with a 400 error. module Truncation - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO = :auto DISABLED = :disabled diff --git a/lib/openai/models/responses/response_created_event.rb b/lib/openai/models/responses/response_created_event.rb index 9a4cc890..2eb916c1 100644 --- a/lib/openai/models/responses/response_created_event.rb +++ b/lib/openai/models/responses/response_created_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseCreatedEvent < OpenAI::BaseModel + class ResponseCreatedEvent < OpenAI::Internal::Type::BaseModel # @!attribute response # The response that was created. # @@ -24,7 +24,7 @@ class ResponseCreatedEvent < OpenAI::BaseModel # # # def initialize(response:, type: :"response.created", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_delete_params.rb b/lib/openai/models/responses/response_delete_params.rb index 311adb57..516661d0 100644 --- a/lib/openai/models/responses/response_delete_params.rb +++ b/lib/openai/models/responses/response_delete_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Responses # @see OpenAI::Resources::Responses#delete - class ResponseDeleteParams < OpenAI::BaseModel + class ResponseDeleteParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -14,7 +14,7 @@ class ResponseDeleteParams < OpenAI::BaseModel # # # def initialize(request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_error.rb b/lib/openai/models/responses/response_error.rb index 465b571a..a42f876f 100644 --- a/lib/openai/models/responses/response_error.rb +++ b/lib/openai/models/responses/response_error.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseError < OpenAI::BaseModel + class ResponseError < OpenAI::Internal::Type::BaseModel # @!attribute code # The error code for the response. # @@ -24,13 +24,13 @@ class ResponseError < OpenAI::BaseModel # # # def initialize(code:, message:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The error code for the response. # # @see OpenAI::Models::Responses::ResponseError#code module Code - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum SERVER_ERROR = :server_error RATE_LIMIT_EXCEEDED = :rate_limit_exceeded diff --git a/lib/openai/models/responses/response_error_event.rb b/lib/openai/models/responses/response_error_event.rb index 057c30e1..5558c8c0 100644 --- a/lib/openai/models/responses/response_error_event.rb +++ b/lib/openai/models/responses/response_error_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseErrorEvent < OpenAI::BaseModel + class ResponseErrorEvent < OpenAI::Internal::Type::BaseModel # @!attribute code # The error code. # @@ -38,7 +38,7 @@ class ResponseErrorEvent < OpenAI::BaseModel # # # def initialize(code:, message:, param:, type: :error, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_failed_event.rb b/lib/openai/models/responses/response_failed_event.rb index 2b4d5471..aa99a457 100644 --- a/lib/openai/models/responses/response_failed_event.rb +++ b/lib/openai/models/responses/response_failed_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseFailedEvent < OpenAI::BaseModel + class ResponseFailedEvent < OpenAI::Internal::Type::BaseModel # @!attribute response # The response that failed. # @@ -24,7 +24,7 @@ class ResponseFailedEvent < OpenAI::BaseModel # # # def initialize(response:, type: :"response.failed", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_file_search_call_completed_event.rb b/lib/openai/models/responses/response_file_search_call_completed_event.rb index 8aaa0427..458771a8 100644 --- a/lib/openai/models/responses/response_file_search_call_completed_event.rb +++ b/lib/openai/models/responses/response_file_search_call_completed_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseFileSearchCallCompletedEvent < OpenAI::BaseModel + class ResponseFileSearchCallCompletedEvent < OpenAI::Internal::Type::BaseModel # @!attribute item_id # The ID of the output item that the file search call is initiated. # @@ -31,7 +31,7 @@ class ResponseFileSearchCallCompletedEvent < OpenAI::BaseModel # # # def initialize(item_id:, output_index:, type: :"response.file_search_call.completed", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_file_search_call_in_progress_event.rb b/lib/openai/models/responses/response_file_search_call_in_progress_event.rb index 09237200..4671a7e6 100644 --- a/lib/openai/models/responses/response_file_search_call_in_progress_event.rb +++ b/lib/openai/models/responses/response_file_search_call_in_progress_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseFileSearchCallInProgressEvent < OpenAI::BaseModel + class ResponseFileSearchCallInProgressEvent < OpenAI::Internal::Type::BaseModel # @!attribute item_id # The ID of the output item that the file search call is initiated. # @@ -31,7 +31,7 @@ class ResponseFileSearchCallInProgressEvent < OpenAI::BaseModel # # # def initialize(item_id:, output_index:, type: :"response.file_search_call.in_progress", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_file_search_call_searching_event.rb b/lib/openai/models/responses/response_file_search_call_searching_event.rb index bc87ce2c..880e33e1 100644 --- a/lib/openai/models/responses/response_file_search_call_searching_event.rb +++ b/lib/openai/models/responses/response_file_search_call_searching_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseFileSearchCallSearchingEvent < OpenAI::BaseModel + class ResponseFileSearchCallSearchingEvent < OpenAI::Internal::Type::BaseModel # @!attribute item_id # The ID of the output item that the file search call is initiated. # @@ -31,7 +31,7 @@ class ResponseFileSearchCallSearchingEvent < OpenAI::BaseModel # # # def initialize(item_id:, output_index:, type: :"response.file_search_call.searching", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_file_search_tool_call.rb b/lib/openai/models/responses/response_file_search_tool_call.rb index 3606686b..dec92ac0 100644 --- a/lib/openai/models/responses/response_file_search_tool_call.rb +++ b/lib/openai/models/responses/response_file_search_tool_call.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseFileSearchToolCall < OpenAI::BaseModel + class ResponseFileSearchToolCall < OpenAI::Internal::Type::BaseModel # @!attribute id # The unique ID of the file search tool call. # @@ -14,7 +14,7 @@ class ResponseFileSearchToolCall < OpenAI::BaseModel # The queries used to search for files. # # @return [Array] - required :queries, OpenAI::ArrayOf[String] + required :queries, OpenAI::Internal::Type::ArrayOf[String] # @!attribute status # The status of the file search tool call. One of `in_progress`, `searching`, @@ -34,7 +34,7 @@ class ResponseFileSearchToolCall < OpenAI::BaseModel # # @return [Array, nil] optional :results, - -> { OpenAI::ArrayOf[OpenAI::Models::Responses::ResponseFileSearchToolCall::Result] }, + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Responses::ResponseFileSearchToolCall::Result] }, nil?: true # @!parse @@ -50,14 +50,14 @@ class ResponseFileSearchToolCall < OpenAI::BaseModel # # # def initialize(id:, queries:, status:, results: nil, type: :file_search_call, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The status of the file search tool call. One of `in_progress`, `searching`, # `incomplete` or `failed`, # # @see OpenAI::Models::Responses::ResponseFileSearchToolCall#status module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS = :in_progress SEARCHING = :searching @@ -72,7 +72,7 @@ module Status # def self.values; end end - class Result < OpenAI::BaseModel + class Result < OpenAI::Internal::Type::BaseModel # @!attribute attributes # Set of 16 key-value pairs that can be attached to an object. This can be useful # for storing additional information about the object in a structured format, and @@ -82,7 +82,7 @@ class Result < OpenAI::BaseModel # # @return [Hash{Symbol=>String, Float, Boolean}, nil] optional :attributes, - -> { OpenAI::HashOf[union: OpenAI::Models::Responses::ResponseFileSearchToolCall::Result::Attribute] }, + -> { OpenAI::Internal::Type::HashOf[union: OpenAI::Models::Responses::ResponseFileSearchToolCall::Result::Attribute] }, nil?: true # @!attribute [r] file_id @@ -134,16 +134,16 @@ class Result < OpenAI::BaseModel # # # def initialize(attributes: nil, file_id: nil, filename: nil, score: nil, text: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void module Attribute - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String variant Float - variant OpenAI::BooleanModel + variant OpenAI::Internal::Type::BooleanModel # @!parse # # @return [Array(String, Float, Boolean)] diff --git a/lib/openai/models/responses/response_format_text_config.rb b/lib/openai/models/responses/response_format_text_config.rb index ee665d54..bdacd77c 100644 --- a/lib/openai/models/responses/response_format_text_config.rb +++ b/lib/openai/models/responses/response_format_text_config.rb @@ -17,7 +17,7 @@ module Responses # ensures the message the model generates is valid JSON. Using `json_schema` is # preferred for models that support it. module ResponseFormatTextConfig - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type diff --git a/lib/openai/models/responses/response_format_text_json_schema_config.rb b/lib/openai/models/responses/response_format_text_json_schema_config.rb index 06e3bf5e..f0404403 100644 --- a/lib/openai/models/responses/response_format_text_json_schema_config.rb +++ b/lib/openai/models/responses/response_format_text_json_schema_config.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseFormatTextJSONSchemaConfig < OpenAI::BaseModel + class ResponseFormatTextJSONSchemaConfig < OpenAI::Internal::Type::BaseModel # @!attribute name # The name of the response format. Must be a-z, A-Z, 0-9, or contain underscores # and dashes, with a maximum length of 64. @@ -16,7 +16,7 @@ class ResponseFormatTextJSONSchemaConfig < OpenAI::BaseModel # to build JSON schemas [here](https://json-schema.org/). # # @return [Hash{Symbol=>Object}] - required :schema, OpenAI::HashOf[OpenAI::Unknown] + required :schema, OpenAI::Internal::Type::HashOf[OpenAI::Internal::Type::Unknown] # @!attribute type # The type of response format being defined. Always `json_schema`. @@ -43,7 +43,7 @@ class ResponseFormatTextJSONSchemaConfig < OpenAI::BaseModel # [Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs). # # @return [Boolean, nil] - optional :strict, OpenAI::BooleanModel, nil?: true + optional :strict, OpenAI::Internal::Type::BooleanModel, nil?: true # @!parse # # JSON Schema response format. Used to generate structured JSON responses. Learn @@ -58,7 +58,7 @@ class ResponseFormatTextJSONSchemaConfig < OpenAI::BaseModel # # # def initialize(name:, schema:, description: nil, strict: nil, type: :json_schema, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_function_call_arguments_delta_event.rb b/lib/openai/models/responses/response_function_call_arguments_delta_event.rb index 51e4b411..6005ba84 100644 --- a/lib/openai/models/responses/response_function_call_arguments_delta_event.rb +++ b/lib/openai/models/responses/response_function_call_arguments_delta_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseFunctionCallArgumentsDeltaEvent < OpenAI::BaseModel + class ResponseFunctionCallArgumentsDeltaEvent < OpenAI::Internal::Type::BaseModel # @!attribute delta # The function-call arguments delta that is added. # @@ -38,7 +38,7 @@ class ResponseFunctionCallArgumentsDeltaEvent < OpenAI::BaseModel # # # def initialize(delta:, item_id:, output_index:, type: :"response.function_call_arguments.delta", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_function_call_arguments_done_event.rb b/lib/openai/models/responses/response_function_call_arguments_done_event.rb index 379a1df5..a9280f46 100644 --- a/lib/openai/models/responses/response_function_call_arguments_done_event.rb +++ b/lib/openai/models/responses/response_function_call_arguments_done_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseFunctionCallArgumentsDoneEvent < OpenAI::BaseModel + class ResponseFunctionCallArgumentsDoneEvent < OpenAI::Internal::Type::BaseModel # @!attribute arguments # The function-call arguments. # @@ -37,7 +37,7 @@ class ResponseFunctionCallArgumentsDoneEvent < OpenAI::BaseModel # # # def initialize(arguments:, item_id:, output_index:, type: :"response.function_call_arguments.done", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_function_tool_call.rb b/lib/openai/models/responses/response_function_tool_call.rb index 92fbd817..61ce02d2 100644 --- a/lib/openai/models/responses/response_function_tool_call.rb +++ b/lib/openai/models/responses/response_function_tool_call.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseFunctionToolCall < OpenAI::BaseModel + class ResponseFunctionToolCall < OpenAI::Internal::Type::BaseModel # @!attribute arguments # A JSON string of the arguments to pass to the function. # @@ -63,14 +63,14 @@ class ResponseFunctionToolCall < OpenAI::BaseModel # # # def initialize(arguments:, call_id:, name:, id: nil, status: nil, type: :function_call, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The status of the item. One of `in_progress`, `completed`, or `incomplete`. # Populated when items are returned via API. # # @see OpenAI::Models::Responses::ResponseFunctionToolCall#status module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS = :in_progress COMPLETED = :completed diff --git a/lib/openai/models/responses/response_function_tool_call_item.rb b/lib/openai/models/responses/response_function_tool_call_item.rb index 9317cee6..8315cb64 100644 --- a/lib/openai/models/responses/response_function_tool_call_item.rb +++ b/lib/openai/models/responses/response_function_tool_call_item.rb @@ -19,7 +19,7 @@ class ResponseFunctionToolCallItem < OpenAI::Models::Responses::ResponseFunction # # # def initialize(id:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_function_tool_call_output_item.rb b/lib/openai/models/responses/response_function_tool_call_output_item.rb index 310a8685..081ef864 100644 --- a/lib/openai/models/responses/response_function_tool_call_output_item.rb +++ b/lib/openai/models/responses/response_function_tool_call_output_item.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseFunctionToolCallOutputItem < OpenAI::BaseModel + class ResponseFunctionToolCallOutputItem < OpenAI::Internal::Type::BaseModel # @!attribute id # The unique ID of the function call tool output. # @@ -48,14 +48,14 @@ class ResponseFunctionToolCallOutputItem < OpenAI::BaseModel # # # def initialize(id:, call_id:, output:, status: nil, type: :function_call_output, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The status of the item. One of `in_progress`, `completed`, or `incomplete`. # Populated when items are returned via API. # # @see OpenAI::Models::Responses::ResponseFunctionToolCallOutputItem#status module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS = :in_progress COMPLETED = :completed diff --git a/lib/openai/models/responses/response_function_web_search.rb b/lib/openai/models/responses/response_function_web_search.rb index 41ecb9f4..c2cdfa12 100644 --- a/lib/openai/models/responses/response_function_web_search.rb +++ b/lib/openai/models/responses/response_function_web_search.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseFunctionWebSearch < OpenAI::BaseModel + class ResponseFunctionWebSearch < OpenAI::Internal::Type::BaseModel # @!attribute id # The unique ID of the web search tool call. # @@ -33,13 +33,13 @@ class ResponseFunctionWebSearch < OpenAI::BaseModel # # # def initialize(id:, status:, type: :web_search_call, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The status of the web search tool call. # # @see OpenAI::Models::Responses::ResponseFunctionWebSearch#status module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS = :in_progress SEARCHING = :searching diff --git a/lib/openai/models/responses/response_in_progress_event.rb b/lib/openai/models/responses/response_in_progress_event.rb index 5f61d802..a26c5659 100644 --- a/lib/openai/models/responses/response_in_progress_event.rb +++ b/lib/openai/models/responses/response_in_progress_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseInProgressEvent < OpenAI::BaseModel + class ResponseInProgressEvent < OpenAI::Internal::Type::BaseModel # @!attribute response # The response that is in progress. # @@ -24,7 +24,7 @@ class ResponseInProgressEvent < OpenAI::BaseModel # # # def initialize(response:, type: :"response.in_progress", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_includable.rb b/lib/openai/models/responses/response_includable.rb index 39763a87..26bd124c 100644 --- a/lib/openai/models/responses/response_includable.rb +++ b/lib/openai/models/responses/response_includable.rb @@ -12,7 +12,7 @@ module Responses # - `computer_call_output.output.image_url`: Include image urls from the computer # call output. module ResponseIncludable - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum FILE_SEARCH_CALL_RESULTS = :"file_search_call.results" MESSAGE_INPUT_IMAGE_IMAGE_URL = :"message.input_image.image_url" diff --git a/lib/openai/models/responses/response_incomplete_event.rb b/lib/openai/models/responses/response_incomplete_event.rb index 8f0f240c..08fb757a 100644 --- a/lib/openai/models/responses/response_incomplete_event.rb +++ b/lib/openai/models/responses/response_incomplete_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseIncompleteEvent < OpenAI::BaseModel + class ResponseIncompleteEvent < OpenAI::Internal::Type::BaseModel # @!attribute response # The response that was incomplete. # @@ -24,7 +24,7 @@ class ResponseIncompleteEvent < OpenAI::BaseModel # # # def initialize(response:, type: :"response.incomplete", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_input.rb b/lib/openai/models/responses/response_input.rb index d8565c95..591bab05 100644 --- a/lib/openai/models/responses/response_input.rb +++ b/lib/openai/models/responses/response_input.rb @@ -3,7 +3,8 @@ module OpenAI module Models module Responses - ResponseInput = OpenAI::ArrayOf[union: -> { OpenAI::Models::Responses::ResponseInputItem }] + ResponseInput = + OpenAI::Internal::Type::ArrayOf[union: -> { OpenAI::Models::Responses::ResponseInputItem }] end end end diff --git a/lib/openai/models/responses/response_input_audio.rb b/lib/openai/models/responses/response_input_audio.rb index e6caa2ed..90dcdee6 100644 --- a/lib/openai/models/responses/response_input_audio.rb +++ b/lib/openai/models/responses/response_input_audio.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseInputAudio < OpenAI::BaseModel + class ResponseInputAudio < OpenAI::Internal::Type::BaseModel # @!attribute data # Base64-encoded audio data. # @@ -31,13 +31,13 @@ class ResponseInputAudio < OpenAI::BaseModel # # # def initialize(data:, format_:, type: :input_audio, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The format of the audio data. Currently supported formats are `mp3` and `wav`. # # @see OpenAI::Models::Responses::ResponseInputAudio#format_ module Format - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum MP3 = :mp3 WAV = :wav diff --git a/lib/openai/models/responses/response_input_content.rb b/lib/openai/models/responses/response_input_content.rb index 41f9a7d7..901a5159 100644 --- a/lib/openai/models/responses/response_input_content.rb +++ b/lib/openai/models/responses/response_input_content.rb @@ -5,7 +5,7 @@ module Models module Responses # A text input to the model. module ResponseInputContent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type diff --git a/lib/openai/models/responses/response_input_file.rb b/lib/openai/models/responses/response_input_file.rb index a6e595f9..0e2a3c73 100644 --- a/lib/openai/models/responses/response_input_file.rb +++ b/lib/openai/models/responses/response_input_file.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseInputFile < OpenAI::BaseModel + class ResponseInputFile < OpenAI::Internal::Type::BaseModel # @!attribute type # The type of the input item. Always `input_file`. # @@ -50,7 +50,7 @@ class ResponseInputFile < OpenAI::BaseModel # # # def initialize(file_data: nil, file_id: nil, filename: nil, type: :input_file, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_input_image.rb b/lib/openai/models/responses/response_input_image.rb index 5c5bb8ce..883f8491 100644 --- a/lib/openai/models/responses/response_input_image.rb +++ b/lib/openai/models/responses/response_input_image.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseInputImage < OpenAI::BaseModel + class ResponseInputImage < OpenAI::Internal::Type::BaseModel # @!attribute detail # The detail level of the image to be sent to the model. One of `high`, `low`, or # `auto`. Defaults to `auto`. @@ -41,14 +41,14 @@ class ResponseInputImage < OpenAI::BaseModel # # # def initialize(detail:, file_id: nil, image_url: nil, type: :input_image, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The detail level of the image to be sent to the model. One of `high`, `low`, or # `auto`. Defaults to `auto`. # # @see OpenAI::Models::Responses::ResponseInputImage#detail module Detail - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum HIGH = :high LOW = :low diff --git a/lib/openai/models/responses/response_input_item.rb b/lib/openai/models/responses/response_input_item.rb index 206650eb..62862e0e 100644 --- a/lib/openai/models/responses/response_input_item.rb +++ b/lib/openai/models/responses/response_input_item.rb @@ -9,7 +9,7 @@ module Responses # `assistant` role are presumed to have been generated by the model in previous # interactions. module ResponseInputItem - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type @@ -57,13 +57,14 @@ module ResponseInputItem # An internal identifier for an item to reference. variant :item_reference, -> { OpenAI::Models::Responses::ResponseInputItem::ItemReference } - class Message < OpenAI::BaseModel + class Message < OpenAI::Internal::Type::BaseModel # @!attribute content # A list of one or many input items to the model, containing different content # types. # # @return [Array] - required :content, -> { OpenAI::ArrayOf[union: OpenAI::Models::Responses::ResponseInputContent] } + required :content, + -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Responses::ResponseInputContent] } # @!attribute role # The role of the message input. One of `user`, `system`, or `developer`. @@ -104,13 +105,13 @@ class Message < OpenAI::BaseModel # # # def initialize(content:, role:, status: nil, type: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The role of the message input. One of `user`, `system`, or `developer`. # # @see OpenAI::Models::Responses::ResponseInputItem::Message#role module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum USER = :user SYSTEM = :system @@ -128,7 +129,7 @@ module Role # # @see OpenAI::Models::Responses::ResponseInputItem::Message#status module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS = :in_progress COMPLETED = :completed @@ -145,7 +146,7 @@ module Status # # @see OpenAI::Models::Responses::ResponseInputItem::Message#type module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum MESSAGE = :message @@ -157,7 +158,7 @@ module Type end end - class ComputerCallOutput < OpenAI::BaseModel + class ComputerCallOutput < OpenAI::Internal::Type::BaseModel # @!attribute call_id # The ID of the computer tool call that produced the output. # @@ -192,7 +193,7 @@ class ComputerCallOutput < OpenAI::BaseModel # # @return [Array, nil] optional :acknowledged_safety_checks, - -> { OpenAI::ArrayOf[OpenAI::Models::Responses::ResponseInputItem::ComputerCallOutput::AcknowledgedSafetyCheck] } + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Responses::ResponseInputItem::ComputerCallOutput::AcknowledgedSafetyCheck] } # @!parse # # @return [Array] @@ -221,9 +222,9 @@ class ComputerCallOutput < OpenAI::BaseModel # # # def initialize(call_id:, output:, id: nil, acknowledged_safety_checks: nil, status: nil, type: :computer_call_output, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class AcknowledgedSafetyCheck < OpenAI::BaseModel + class AcknowledgedSafetyCheck < OpenAI::Internal::Type::BaseModel # @!attribute id # The ID of the pending safety check. # @@ -251,7 +252,7 @@ class AcknowledgedSafetyCheck < OpenAI::BaseModel # # # def initialize(id:, code:, message:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # The status of the message input. One of `in_progress`, `completed`, or @@ -259,7 +260,7 @@ class AcknowledgedSafetyCheck < OpenAI::BaseModel # # @see OpenAI::Models::Responses::ResponseInputItem::ComputerCallOutput#status module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS = :in_progress COMPLETED = :completed @@ -273,7 +274,7 @@ module Status end end - class FunctionCallOutput < OpenAI::BaseModel + class FunctionCallOutput < OpenAI::Internal::Type::BaseModel # @!attribute call_id # The unique ID of the function tool call generated by the model. # @@ -325,14 +326,14 @@ class FunctionCallOutput < OpenAI::BaseModel # # # def initialize(call_id:, output:, id: nil, status: nil, type: :function_call_output, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The status of the item. One of `in_progress`, `completed`, or `incomplete`. # Populated when items are returned via API. # # @see OpenAI::Models::Responses::ResponseInputItem::FunctionCallOutput#status module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS = :in_progress COMPLETED = :completed @@ -346,7 +347,7 @@ module Status end end - class ItemReference < OpenAI::BaseModel + class ItemReference < OpenAI::Internal::Type::BaseModel # @!attribute id # The ID of the item to reference. # @@ -367,7 +368,7 @@ class ItemReference < OpenAI::BaseModel # # # def initialize(id:, type: :item_reference, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @!parse diff --git a/lib/openai/models/responses/response_input_message_content_list.rb b/lib/openai/models/responses/response_input_message_content_list.rb index 98901e68..a1362365 100644 --- a/lib/openai/models/responses/response_input_message_content_list.rb +++ b/lib/openai/models/responses/response_input_message_content_list.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Responses ResponseInputMessageContentList = - OpenAI::ArrayOf[union: -> { OpenAI::Models::Responses::ResponseInputContent }] + OpenAI::Internal::Type::ArrayOf[union: -> { OpenAI::Models::Responses::ResponseInputContent }] end end end diff --git a/lib/openai/models/responses/response_input_message_item.rb b/lib/openai/models/responses/response_input_message_item.rb index 65bfbf76..3bb170a9 100644 --- a/lib/openai/models/responses/response_input_message_item.rb +++ b/lib/openai/models/responses/response_input_message_item.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseInputMessageItem < OpenAI::BaseModel + class ResponseInputMessageItem < OpenAI::Internal::Type::BaseModel # @!attribute id # The unique ID of the message input. # @@ -15,7 +15,8 @@ class ResponseInputMessageItem < OpenAI::BaseModel # types. # # @return [Array] - required :content, -> { OpenAI::ArrayOf[union: OpenAI::Models::Responses::ResponseInputContent] } + required :content, + -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Responses::ResponseInputContent] } # @!attribute role # The role of the message input. One of `user`, `system`, or `developer`. @@ -53,13 +54,13 @@ class ResponseInputMessageItem < OpenAI::BaseModel # # # def initialize(id:, content:, role:, status: nil, type: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The role of the message input. One of `user`, `system`, or `developer`. # # @see OpenAI::Models::Responses::ResponseInputMessageItem#role module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum USER = :user SYSTEM = :system @@ -77,7 +78,7 @@ module Role # # @see OpenAI::Models::Responses::ResponseInputMessageItem#status module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS = :in_progress COMPLETED = :completed @@ -94,7 +95,7 @@ module Status # # @see OpenAI::Models::Responses::ResponseInputMessageItem#type module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum MESSAGE = :message diff --git a/lib/openai/models/responses/response_input_text.rb b/lib/openai/models/responses/response_input_text.rb index c38be5e5..d8ed6f2c 100644 --- a/lib/openai/models/responses/response_input_text.rb +++ b/lib/openai/models/responses/response_input_text.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseInputText < OpenAI::BaseModel + class ResponseInputText < OpenAI::Internal::Type::BaseModel # @!attribute text # The text input to the model. # @@ -24,7 +24,7 @@ class ResponseInputText < OpenAI::BaseModel # # # def initialize(text:, type: :input_text, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_item.rb b/lib/openai/models/responses/response_item.rb index ca1c9ee7..3fe0074e 100644 --- a/lib/openai/models/responses/response_item.rb +++ b/lib/openai/models/responses/response_item.rb @@ -5,7 +5,7 @@ module Models module Responses # Content item used to generate a response. module ResponseItem - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type diff --git a/lib/openai/models/responses/response_item_list.rb b/lib/openai/models/responses/response_item_list.rb index 34ca03ad..5c05dab2 100644 --- a/lib/openai/models/responses/response_item_list.rb +++ b/lib/openai/models/responses/response_item_list.rb @@ -3,12 +3,12 @@ module OpenAI module Models module Responses - class ResponseItemList < OpenAI::BaseModel + class ResponseItemList < OpenAI::Internal::Type::BaseModel # @!attribute data # A list of items used to generate this response. # # @return [Array] - required :data, -> { OpenAI::ArrayOf[union: OpenAI::Models::Responses::ResponseItem] } + required :data, -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Responses::ResponseItem] } # @!attribute first_id # The ID of the first item in the list. @@ -20,7 +20,7 @@ class ResponseItemList < OpenAI::BaseModel # Whether there are more items available. # # @return [Boolean] - required :has_more, OpenAI::BooleanModel + required :has_more, OpenAI::Internal::Type::BooleanModel # @!attribute last_id # The ID of the last item in the list. @@ -45,7 +45,7 @@ class ResponseItemList < OpenAI::BaseModel # # # def initialize(data:, first_id:, has_more:, last_id:, object: :list, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end diff --git a/lib/openai/models/responses/response_output_audio.rb b/lib/openai/models/responses/response_output_audio.rb index da147435..d8fb8c61 100644 --- a/lib/openai/models/responses/response_output_audio.rb +++ b/lib/openai/models/responses/response_output_audio.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseOutputAudio < OpenAI::BaseModel + class ResponseOutputAudio < OpenAI::Internal::Type::BaseModel # @!attribute data # Base64-encoded audio data from the model. # @@ -31,7 +31,7 @@ class ResponseOutputAudio < OpenAI::BaseModel # # # def initialize(data:, transcript:, type: :output_audio, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_output_item.rb b/lib/openai/models/responses/response_output_item.rb index 2666a922..5751bb93 100644 --- a/lib/openai/models/responses/response_output_item.rb +++ b/lib/openai/models/responses/response_output_item.rb @@ -5,7 +5,7 @@ module Models module Responses # An output message from the model. module ResponseOutputItem - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type diff --git a/lib/openai/models/responses/response_output_item_added_event.rb b/lib/openai/models/responses/response_output_item_added_event.rb index 9885b0a9..6dfbd4d9 100644 --- a/lib/openai/models/responses/response_output_item_added_event.rb +++ b/lib/openai/models/responses/response_output_item_added_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseOutputItemAddedEvent < OpenAI::BaseModel + class ResponseOutputItemAddedEvent < OpenAI::Internal::Type::BaseModel # @!attribute item # The output item that was added. # @@ -31,7 +31,7 @@ class ResponseOutputItemAddedEvent < OpenAI::BaseModel # # # def initialize(item:, output_index:, type: :"response.output_item.added", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_output_item_done_event.rb b/lib/openai/models/responses/response_output_item_done_event.rb index 6840bbd7..904c8eae 100644 --- a/lib/openai/models/responses/response_output_item_done_event.rb +++ b/lib/openai/models/responses/response_output_item_done_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseOutputItemDoneEvent < OpenAI::BaseModel + class ResponseOutputItemDoneEvent < OpenAI::Internal::Type::BaseModel # @!attribute item # The output item that was marked done. # @@ -31,7 +31,7 @@ class ResponseOutputItemDoneEvent < OpenAI::BaseModel # # # def initialize(item:, output_index:, type: :"response.output_item.done", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_output_message.rb b/lib/openai/models/responses/response_output_message.rb index 373721b0..272c8ab1 100644 --- a/lib/openai/models/responses/response_output_message.rb +++ b/lib/openai/models/responses/response_output_message.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseOutputMessage < OpenAI::BaseModel + class ResponseOutputMessage < OpenAI::Internal::Type::BaseModel # @!attribute id # The unique ID of the output message. # @@ -15,7 +15,7 @@ class ResponseOutputMessage < OpenAI::BaseModel # # @return [Array] required :content, - -> { OpenAI::ArrayOf[union: OpenAI::Models::Responses::ResponseOutputMessage::Content] } + -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Responses::ResponseOutputMessage::Content] } # @!attribute role # The role of the output message. Always `assistant`. @@ -47,11 +47,11 @@ class ResponseOutputMessage < OpenAI::BaseModel # # # def initialize(id:, content:, status:, role: :assistant, type: :message, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # A text output from the model. module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type @@ -71,7 +71,7 @@ module Content # # @see OpenAI::Models::Responses::ResponseOutputMessage#status module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS = :in_progress COMPLETED = :completed diff --git a/lib/openai/models/responses/response_output_refusal.rb b/lib/openai/models/responses/response_output_refusal.rb index bd44ae1e..0e050ce2 100644 --- a/lib/openai/models/responses/response_output_refusal.rb +++ b/lib/openai/models/responses/response_output_refusal.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseOutputRefusal < OpenAI::BaseModel + class ResponseOutputRefusal < OpenAI::Internal::Type::BaseModel # @!attribute refusal # The refusal explanationfrom the model. # @@ -24,7 +24,7 @@ class ResponseOutputRefusal < OpenAI::BaseModel # # # def initialize(refusal:, type: :refusal, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_output_text.rb b/lib/openai/models/responses/response_output_text.rb index e6c46e68..9766a767 100644 --- a/lib/openai/models/responses/response_output_text.rb +++ b/lib/openai/models/responses/response_output_text.rb @@ -3,13 +3,13 @@ module OpenAI module Models module Responses - class ResponseOutputText < OpenAI::BaseModel + class ResponseOutputText < OpenAI::Internal::Type::BaseModel # @!attribute annotations # The annotations of the text output. # # @return [Array] required :annotations, - -> { OpenAI::ArrayOf[union: OpenAI::Models::Responses::ResponseOutputText::Annotation] } + -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Responses::ResponseOutputText::Annotation] } # @!attribute text # The text output from the model. @@ -32,11 +32,11 @@ class ResponseOutputText < OpenAI::BaseModel # # # def initialize(annotations:, text:, type: :output_text, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # A citation to a file. module Annotation - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type @@ -49,7 +49,7 @@ module Annotation # A path to a file. variant :file_path, -> { OpenAI::Models::Responses::ResponseOutputText::Annotation::FilePath } - class FileCitation < OpenAI::BaseModel + class FileCitation < OpenAI::Internal::Type::BaseModel # @!attribute file_id # The ID of the file. # @@ -77,10 +77,10 @@ class FileCitation < OpenAI::BaseModel # # # def initialize(file_id:, index:, type: :file_citation, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class URLCitation < OpenAI::BaseModel + class URLCitation < OpenAI::Internal::Type::BaseModel # @!attribute end_index # The index of the last character of the URL citation in the message. # @@ -122,10 +122,10 @@ class URLCitation < OpenAI::BaseModel # # # def initialize(end_index:, start_index:, title:, url:, type: :url_citation, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class FilePath < OpenAI::BaseModel + class FilePath < OpenAI::Internal::Type::BaseModel # @!attribute file_id # The ID of the file. # @@ -153,7 +153,7 @@ class FilePath < OpenAI::BaseModel # # # def initialize(file_id:, index:, type: :file_path, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @!parse diff --git a/lib/openai/models/responses/response_reasoning_item.rb b/lib/openai/models/responses/response_reasoning_item.rb index e5539f3d..fa4d428c 100644 --- a/lib/openai/models/responses/response_reasoning_item.rb +++ b/lib/openai/models/responses/response_reasoning_item.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseReasoningItem < OpenAI::BaseModel + class ResponseReasoningItem < OpenAI::Internal::Type::BaseModel # @!attribute id # The unique identifier of the reasoning content. # @@ -14,7 +14,8 @@ class ResponseReasoningItem < OpenAI::BaseModel # Reasoning text contents. # # @return [Array] - required :summary, -> { OpenAI::ArrayOf[OpenAI::Models::Responses::ResponseReasoningItem::Summary] } + required :summary, + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Responses::ResponseReasoningItem::Summary] } # @!attribute type # The type of the object. Always `reasoning`. @@ -44,9 +45,9 @@ class ResponseReasoningItem < OpenAI::BaseModel # # # def initialize(id:, summary:, status: nil, type: :reasoning, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class Summary < OpenAI::BaseModel + class Summary < OpenAI::Internal::Type::BaseModel # @!attribute text # A short summary of the reasoning used by the model when generating the response. # @@ -65,7 +66,7 @@ class Summary < OpenAI::BaseModel # # # def initialize(text:, type: :summary_text, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # The status of the item. One of `in_progress`, `completed`, or `incomplete`. @@ -73,7 +74,7 @@ class Summary < OpenAI::BaseModel # # @see OpenAI::Models::Responses::ResponseReasoningItem#status module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS = :in_progress COMPLETED = :completed diff --git a/lib/openai/models/responses/response_refusal_delta_event.rb b/lib/openai/models/responses/response_refusal_delta_event.rb index b09b8ef2..35247a7d 100644 --- a/lib/openai/models/responses/response_refusal_delta_event.rb +++ b/lib/openai/models/responses/response_refusal_delta_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseRefusalDeltaEvent < OpenAI::BaseModel + class ResponseRefusalDeltaEvent < OpenAI::Internal::Type::BaseModel # @!attribute content_index # The index of the content part that the refusal text is added to. # @@ -45,7 +45,7 @@ class ResponseRefusalDeltaEvent < OpenAI::BaseModel # # # def initialize(content_index:, delta:, item_id:, output_index:, type: :"response.refusal.delta", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_refusal_done_event.rb b/lib/openai/models/responses/response_refusal_done_event.rb index 662705f7..7f6cd16b 100644 --- a/lib/openai/models/responses/response_refusal_done_event.rb +++ b/lib/openai/models/responses/response_refusal_done_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseRefusalDoneEvent < OpenAI::BaseModel + class ResponseRefusalDoneEvent < OpenAI::Internal::Type::BaseModel # @!attribute content_index # The index of the content part that the refusal text is finalized. # @@ -45,7 +45,7 @@ class ResponseRefusalDoneEvent < OpenAI::BaseModel # # # def initialize(content_index:, item_id:, output_index:, refusal:, type: :"response.refusal.done", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_retrieve_params.rb b/lib/openai/models/responses/response_retrieve_params.rb index ea7d0882..eaef643f 100644 --- a/lib/openai/models/responses/response_retrieve_params.rb +++ b/lib/openai/models/responses/response_retrieve_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Responses # @see OpenAI::Resources::Responses#retrieve - class ResponseRetrieveParams < OpenAI::BaseModel + class ResponseRetrieveParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -14,7 +14,8 @@ class ResponseRetrieveParams < OpenAI::BaseModel # Response creation above for more information. # # @return [Array, nil] - optional :include, -> { OpenAI::ArrayOf[enum: OpenAI::Models::Responses::ResponseIncludable] } + optional :include, + -> { OpenAI::Internal::Type::ArrayOf[enum: OpenAI::Models::Responses::ResponseIncludable] } # @!parse # # @return [Array] @@ -26,7 +27,7 @@ class ResponseRetrieveParams < OpenAI::BaseModel # # # def initialize(include: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_status.rb b/lib/openai/models/responses/response_status.rb index 31187039..15876663 100644 --- a/lib/openai/models/responses/response_status.rb +++ b/lib/openai/models/responses/response_status.rb @@ -6,7 +6,7 @@ module Responses # The status of the response generation. One of `completed`, `failed`, # `in_progress`, or `incomplete`. module ResponseStatus - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum COMPLETED = :completed FAILED = :failed diff --git a/lib/openai/models/responses/response_stream_event.rb b/lib/openai/models/responses/response_stream_event.rb index 5dddc41f..db86c410 100644 --- a/lib/openai/models/responses/response_stream_event.rb +++ b/lib/openai/models/responses/response_stream_event.rb @@ -5,7 +5,7 @@ module Models module Responses # Emitted when there is a partial audio response. module ResponseStreamEvent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type diff --git a/lib/openai/models/responses/response_text_annotation_delta_event.rb b/lib/openai/models/responses/response_text_annotation_delta_event.rb index c6291cf6..bc38658a 100644 --- a/lib/openai/models/responses/response_text_annotation_delta_event.rb +++ b/lib/openai/models/responses/response_text_annotation_delta_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseTextAnnotationDeltaEvent < OpenAI::BaseModel + class ResponseTextAnnotationDeltaEvent < OpenAI::Internal::Type::BaseModel # @!attribute annotation # A citation to a file. # @@ -63,13 +63,13 @@ class ResponseTextAnnotationDeltaEvent < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # A citation to a file. # # @see OpenAI::Models::Responses::ResponseTextAnnotationDeltaEvent#annotation module Annotation - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type @@ -85,7 +85,7 @@ module Annotation variant :file_path, -> { OpenAI::Models::Responses::ResponseTextAnnotationDeltaEvent::Annotation::FilePath } - class FileCitation < OpenAI::BaseModel + class FileCitation < OpenAI::Internal::Type::BaseModel # @!attribute file_id # The ID of the file. # @@ -113,10 +113,10 @@ class FileCitation < OpenAI::BaseModel # # # def initialize(file_id:, index:, type: :file_citation, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class URLCitation < OpenAI::BaseModel + class URLCitation < OpenAI::Internal::Type::BaseModel # @!attribute end_index # The index of the last character of the URL citation in the message. # @@ -158,10 +158,10 @@ class URLCitation < OpenAI::BaseModel # # # def initialize(end_index:, start_index:, title:, url:, type: :url_citation, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end - class FilePath < OpenAI::BaseModel + class FilePath < OpenAI::Internal::Type::BaseModel # @!attribute file_id # The ID of the file. # @@ -189,7 +189,7 @@ class FilePath < OpenAI::BaseModel # # # def initialize(file_id:, index:, type: :file_path, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @!parse diff --git a/lib/openai/models/responses/response_text_config.rb b/lib/openai/models/responses/response_text_config.rb index 77a2d9b0..65f31b36 100644 --- a/lib/openai/models/responses/response_text_config.rb +++ b/lib/openai/models/responses/response_text_config.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseTextConfig < OpenAI::BaseModel + class ResponseTextConfig < OpenAI::Internal::Type::BaseModel # @!attribute [r] format_ # An object specifying the format that the model must output. # @@ -37,7 +37,7 @@ class ResponseTextConfig < OpenAI::BaseModel # # # def initialize(format_: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_text_delta_event.rb b/lib/openai/models/responses/response_text_delta_event.rb index 5dee2ee1..8b6c4b75 100644 --- a/lib/openai/models/responses/response_text_delta_event.rb +++ b/lib/openai/models/responses/response_text_delta_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseTextDeltaEvent < OpenAI::BaseModel + class ResponseTextDeltaEvent < OpenAI::Internal::Type::BaseModel # @!attribute content_index # The index of the content part that the text delta was added to. # @@ -45,7 +45,7 @@ class ResponseTextDeltaEvent < OpenAI::BaseModel # # # def initialize(content_index:, delta:, item_id:, output_index:, type: :"response.output_text.delta", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_text_done_event.rb b/lib/openai/models/responses/response_text_done_event.rb index cd3d0faf..45a3267a 100644 --- a/lib/openai/models/responses/response_text_done_event.rb +++ b/lib/openai/models/responses/response_text_done_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseTextDoneEvent < OpenAI::BaseModel + class ResponseTextDoneEvent < OpenAI::Internal::Type::BaseModel # @!attribute content_index # The index of the content part that the text content is finalized. # @@ -45,7 +45,7 @@ class ResponseTextDoneEvent < OpenAI::BaseModel # # # def initialize(content_index:, item_id:, output_index:, text:, type: :"response.output_text.done", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_usage.rb b/lib/openai/models/responses/response_usage.rb index 316d5b3f..8d6bee6d 100644 --- a/lib/openai/models/responses/response_usage.rb +++ b/lib/openai/models/responses/response_usage.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseUsage < OpenAI::BaseModel + class ResponseUsage < OpenAI::Internal::Type::BaseModel # @!attribute input_tokens # The number of input tokens. # @@ -46,10 +46,10 @@ class ResponseUsage < OpenAI::BaseModel # # # def initialize(input_tokens:, input_tokens_details:, output_tokens:, output_tokens_details:, total_tokens:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::Responses::ResponseUsage#input_tokens_details - class InputTokensDetails < OpenAI::BaseModel + class InputTokensDetails < OpenAI::Internal::Type::BaseModel # @!attribute cached_tokens # The number of tokens that were retrieved from the cache. # [More on prompt caching](https://platform.openai.com/docs/guides/prompt-caching). @@ -64,11 +64,11 @@ class InputTokensDetails < OpenAI::BaseModel # # # def initialize(cached_tokens:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # @see OpenAI::Models::Responses::ResponseUsage#output_tokens_details - class OutputTokensDetails < OpenAI::BaseModel + class OutputTokensDetails < OpenAI::Internal::Type::BaseModel # @!attribute reasoning_tokens # The number of reasoning tokens. # @@ -82,7 +82,7 @@ class OutputTokensDetails < OpenAI::BaseModel # # # def initialize(reasoning_tokens:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_web_search_call_completed_event.rb b/lib/openai/models/responses/response_web_search_call_completed_event.rb index 08386de1..70e0dc19 100644 --- a/lib/openai/models/responses/response_web_search_call_completed_event.rb +++ b/lib/openai/models/responses/response_web_search_call_completed_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseWebSearchCallCompletedEvent < OpenAI::BaseModel + class ResponseWebSearchCallCompletedEvent < OpenAI::Internal::Type::BaseModel # @!attribute item_id # Unique ID for the output item associated with the web search call. # @@ -31,7 +31,7 @@ class ResponseWebSearchCallCompletedEvent < OpenAI::BaseModel # # # def initialize(item_id:, output_index:, type: :"response.web_search_call.completed", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_web_search_call_in_progress_event.rb b/lib/openai/models/responses/response_web_search_call_in_progress_event.rb index 0770e80c..fc8f006a 100644 --- a/lib/openai/models/responses/response_web_search_call_in_progress_event.rb +++ b/lib/openai/models/responses/response_web_search_call_in_progress_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseWebSearchCallInProgressEvent < OpenAI::BaseModel + class ResponseWebSearchCallInProgressEvent < OpenAI::Internal::Type::BaseModel # @!attribute item_id # Unique ID for the output item associated with the web search call. # @@ -31,7 +31,7 @@ class ResponseWebSearchCallInProgressEvent < OpenAI::BaseModel # # # def initialize(item_id:, output_index:, type: :"response.web_search_call.in_progress", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/response_web_search_call_searching_event.rb b/lib/openai/models/responses/response_web_search_call_searching_event.rb index 6d037de9..39d6ae9f 100644 --- a/lib/openai/models/responses/response_web_search_call_searching_event.rb +++ b/lib/openai/models/responses/response_web_search_call_searching_event.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseWebSearchCallSearchingEvent < OpenAI::BaseModel + class ResponseWebSearchCallSearchingEvent < OpenAI::Internal::Type::BaseModel # @!attribute item_id # Unique ID for the output item associated with the web search call. # @@ -31,7 +31,7 @@ class ResponseWebSearchCallSearchingEvent < OpenAI::BaseModel # # # def initialize(item_id:, output_index:, type: :"response.web_search_call.searching", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/tool.rb b/lib/openai/models/responses/tool.rb index 3072312f..9c696f1f 100644 --- a/lib/openai/models/responses/tool.rb +++ b/lib/openai/models/responses/tool.rb @@ -7,7 +7,7 @@ module Responses # the # [file search tool](https://platform.openai.com/docs/guides/tools-file-search). module Tool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type diff --git a/lib/openai/models/responses/tool_choice_function.rb b/lib/openai/models/responses/tool_choice_function.rb index 08f9d77e..ca4d89b6 100644 --- a/lib/openai/models/responses/tool_choice_function.rb +++ b/lib/openai/models/responses/tool_choice_function.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ToolChoiceFunction < OpenAI::BaseModel + class ToolChoiceFunction < OpenAI::Internal::Type::BaseModel # @!attribute name # The name of the function to call. # @@ -24,7 +24,7 @@ class ToolChoiceFunction < OpenAI::BaseModel # # # def initialize(name:, type: :function, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses/tool_choice_options.rb b/lib/openai/models/responses/tool_choice_options.rb index 00c16a94..7b08e8c7 100644 --- a/lib/openai/models/responses/tool_choice_options.rb +++ b/lib/openai/models/responses/tool_choice_options.rb @@ -12,7 +12,7 @@ module Responses # # `required` means the model must call one or more tools. module ToolChoiceOptions - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum NONE = :none AUTO = :auto diff --git a/lib/openai/models/responses/tool_choice_types.rb b/lib/openai/models/responses/tool_choice_types.rb index cddbcc56..fe1d606d 100644 --- a/lib/openai/models/responses/tool_choice_types.rb +++ b/lib/openai/models/responses/tool_choice_types.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ToolChoiceTypes < OpenAI::BaseModel + class ToolChoiceTypes < OpenAI::Internal::Type::BaseModel # @!attribute type # The type of hosted tool the model should to use. Learn more about # [built-in tools](https://platform.openai.com/docs/guides/tools). @@ -25,7 +25,7 @@ class ToolChoiceTypes < OpenAI::BaseModel # # # def initialize(type:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The type of hosted tool the model should to use. Learn more about # [built-in tools](https://platform.openai.com/docs/guides/tools). @@ -38,7 +38,7 @@ class ToolChoiceTypes < OpenAI::BaseModel # # @see OpenAI::Models::Responses::ToolChoiceTypes#type module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum FILE_SEARCH = :file_search WEB_SEARCH_PREVIEW = :web_search_preview diff --git a/lib/openai/models/responses/web_search_tool.rb b/lib/openai/models/responses/web_search_tool.rb index 58f833d7..504bcab1 100644 --- a/lib/openai/models/responses/web_search_tool.rb +++ b/lib/openai/models/responses/web_search_tool.rb @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class WebSearchTool < OpenAI::BaseModel + class WebSearchTool < OpenAI::Internal::Type::BaseModel # @!attribute type # The type of the web search tool. One of: # @@ -40,7 +40,7 @@ class WebSearchTool < OpenAI::BaseModel # # # def initialize(type:, search_context_size: nil, user_location: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The type of the web search tool. One of: # @@ -49,7 +49,7 @@ class WebSearchTool < OpenAI::BaseModel # # @see OpenAI::Models::Responses::WebSearchTool#type module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum WEB_SEARCH_PREVIEW = :web_search_preview WEB_SEARCH_PREVIEW_2025_03_11 = :web_search_preview_2025_03_11 @@ -66,7 +66,7 @@ module Type # # @see OpenAI::Models::Responses::WebSearchTool#search_context_size module SearchContextSize - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum LOW = :low MEDIUM = :medium @@ -80,7 +80,7 @@ module SearchContextSize end # @see OpenAI::Models::Responses::WebSearchTool#user_location - class UserLocation < OpenAI::BaseModel + class UserLocation < OpenAI::Internal::Type::BaseModel # @!attribute type # The type of location approximation. Always `approximate`. # @@ -138,7 +138,7 @@ class UserLocation < OpenAI::BaseModel # # # def initialize(city: nil, country: nil, region: nil, timezone: nil, type: :approximate, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/responses_model.rb b/lib/openai/models/responses_model.rb index a7995455..154ea09f 100644 --- a/lib/openai/models/responses_model.rb +++ b/lib/openai/models/responses_model.rb @@ -3,7 +3,7 @@ module OpenAI module Models module ResponsesModel - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String diff --git a/lib/openai/models/static_file_chunking_strategy.rb b/lib/openai/models/static_file_chunking_strategy.rb index 730e687e..9f466cc0 100644 --- a/lib/openai/models/static_file_chunking_strategy.rb +++ b/lib/openai/models/static_file_chunking_strategy.rb @@ -2,7 +2,7 @@ module OpenAI module Models - class StaticFileChunkingStrategy < OpenAI::BaseModel + class StaticFileChunkingStrategy < OpenAI::Internal::Type::BaseModel # @!attribute chunk_overlap_tokens # The number of tokens that overlap between chunks. The default value is `400`. # @@ -24,7 +24,7 @@ class StaticFileChunkingStrategy < OpenAI::BaseModel # # # def initialize(chunk_overlap_tokens:, max_chunk_size_tokens:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/static_file_chunking_strategy_object.rb b/lib/openai/models/static_file_chunking_strategy_object.rb index be6299c5..1655679f 100644 --- a/lib/openai/models/static_file_chunking_strategy_object.rb +++ b/lib/openai/models/static_file_chunking_strategy_object.rb @@ -2,7 +2,7 @@ module OpenAI module Models - class StaticFileChunkingStrategyObject < OpenAI::BaseModel + class StaticFileChunkingStrategyObject < OpenAI::Internal::Type::BaseModel # @!attribute static # # @return [OpenAI::Models::StaticFileChunkingStrategy] @@ -20,7 +20,7 @@ class StaticFileChunkingStrategyObject < OpenAI::BaseModel # # # def initialize(static:, type: :static, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/static_file_chunking_strategy_object_param.rb b/lib/openai/models/static_file_chunking_strategy_object_param.rb index 21b6d0ac..f64fff68 100644 --- a/lib/openai/models/static_file_chunking_strategy_object_param.rb +++ b/lib/openai/models/static_file_chunking_strategy_object_param.rb @@ -2,7 +2,7 @@ module OpenAI module Models - class StaticFileChunkingStrategyObjectParam < OpenAI::BaseModel + class StaticFileChunkingStrategyObjectParam < OpenAI::Internal::Type::BaseModel # @!attribute static # # @return [OpenAI::Models::StaticFileChunkingStrategy] @@ -22,7 +22,7 @@ class StaticFileChunkingStrategyObjectParam < OpenAI::BaseModel # # # def initialize(static:, type: :static, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/upload.rb b/lib/openai/models/upload.rb index a905f9ff..d0da7a8f 100644 --- a/lib/openai/models/upload.rb +++ b/lib/openai/models/upload.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Uploads#create - class Upload < OpenAI::BaseModel + class Upload < OpenAI::Internal::Type::BaseModel # @!attribute id # The Upload unique identifier, which can be referenced in API endpoints. # @@ -75,13 +75,13 @@ class Upload < OpenAI::BaseModel # # # def initialize(id:, bytes:, created_at:, expires_at:, filename:, purpose:, status:, file: nil, object: :upload, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The status of the Upload. # # @see OpenAI::Models::Upload#status module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum PENDING = :pending COMPLETED = :completed diff --git a/lib/openai/models/upload_cancel_params.rb b/lib/openai/models/upload_cancel_params.rb index 5602bb3f..7c44f8c9 100644 --- a/lib/openai/models/upload_cancel_params.rb +++ b/lib/openai/models/upload_cancel_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Uploads#cancel - class UploadCancelParams < OpenAI::BaseModel + class UploadCancelParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -13,7 +13,7 @@ class UploadCancelParams < OpenAI::BaseModel # # # def initialize(request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/upload_complete_params.rb b/lib/openai/models/upload_complete_params.rb index 4b5b8cc3..378f2269 100644 --- a/lib/openai/models/upload_complete_params.rb +++ b/lib/openai/models/upload_complete_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Uploads#complete - class UploadCompleteParams < OpenAI::BaseModel + class UploadCompleteParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -12,7 +12,7 @@ class UploadCompleteParams < OpenAI::BaseModel # The ordered list of Part IDs. # # @return [Array] - required :part_ids, OpenAI::ArrayOf[String] + required :part_ids, OpenAI::Internal::Type::ArrayOf[String] # @!attribute [r] md5 # The optional md5 checksum for the file contents to verify if the bytes uploaded @@ -32,7 +32,7 @@ class UploadCompleteParams < OpenAI::BaseModel # # # def initialize(part_ids:, md5: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/upload_create_params.rb b/lib/openai/models/upload_create_params.rb index b77c4093..9b199b9d 100644 --- a/lib/openai/models/upload_create_params.rb +++ b/lib/openai/models/upload_create_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::Uploads#create - class UploadCreateParams < OpenAI::BaseModel + class UploadCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -47,7 +47,7 @@ class UploadCreateParams < OpenAI::BaseModel # # # def initialize(bytes:, filename:, mime_type:, purpose:, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/uploads/part_create_params.rb b/lib/openai/models/uploads/part_create_params.rb index 12a0fcaf..ff3805ca 100644 --- a/lib/openai/models/uploads/part_create_params.rb +++ b/lib/openai/models/uploads/part_create_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Uploads # @see OpenAI::Resources::Uploads::Parts#create - class PartCreateParams < OpenAI::BaseModel + class PartCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -21,7 +21,7 @@ class PartCreateParams < OpenAI::BaseModel # # # def initialize(data:, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/uploads/upload_part.rb b/lib/openai/models/uploads/upload_part.rb index e0566eb9..4f839461 100644 --- a/lib/openai/models/uploads/upload_part.rb +++ b/lib/openai/models/uploads/upload_part.rb @@ -4,7 +4,7 @@ module OpenAI module Models module Uploads # @see OpenAI::Resources::Uploads::Parts#create - class UploadPart < OpenAI::BaseModel + class UploadPart < OpenAI::Internal::Type::BaseModel # @!attribute id # The upload Part unique identifier, which can be referenced in API endpoints. # @@ -39,7 +39,7 @@ class UploadPart < OpenAI::BaseModel # # # def initialize(id:, created_at:, upload_id:, object: :"upload.part", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end diff --git a/lib/openai/models/vector_store.rb b/lib/openai/models/vector_store.rb index 29391d3a..4b214ef5 100644 --- a/lib/openai/models/vector_store.rb +++ b/lib/openai/models/vector_store.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::VectorStores#create - class VectorStore < OpenAI::BaseModel + class VectorStore < OpenAI::Internal::Type::BaseModel # @!attribute id # The identifier, which can be referenced in API endpoints. # @@ -36,7 +36,7 @@ class VectorStore < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - required :metadata, OpenAI::HashOf[String], nil?: true + required :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!attribute name # The name of the vector store. @@ -113,10 +113,10 @@ class VectorStore < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::VectorStore#file_counts - class FileCounts < OpenAI::BaseModel + class FileCounts < OpenAI::Internal::Type::BaseModel # @!attribute cancelled # The number of files that were cancelled. # @@ -156,7 +156,7 @@ class FileCounts < OpenAI::BaseModel # # # def initialize(cancelled:, completed:, failed:, in_progress:, total:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # The status of the vector store, which can be either `expired`, `in_progress`, or @@ -165,7 +165,7 @@ class FileCounts < OpenAI::BaseModel # # @see OpenAI::Models::VectorStore#status module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum EXPIRED = :expired IN_PROGRESS = :in_progress @@ -179,7 +179,7 @@ module Status end # @see OpenAI::Models::VectorStore#expires_after - class ExpiresAfter < OpenAI::BaseModel + class ExpiresAfter < OpenAI::Internal::Type::BaseModel # @!attribute anchor # Anchor timestamp after which the expiration policy applies. Supported anchors: # `last_active_at`. @@ -201,7 +201,7 @@ class ExpiresAfter < OpenAI::BaseModel # # # def initialize(days:, anchor: :last_active_at, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/vector_store_create_params.rb b/lib/openai/models/vector_store_create_params.rb index a1e61412..42739728 100644 --- a/lib/openai/models/vector_store_create_params.rb +++ b/lib/openai/models/vector_store_create_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::VectorStores#create - class VectorStoreCreateParams < OpenAI::BaseModel + class VectorStoreCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -35,7 +35,7 @@ class VectorStoreCreateParams < OpenAI::BaseModel # files. # # @return [Array, nil] - optional :file_ids, OpenAI::ArrayOf[String] + optional :file_ids, OpenAI::Internal::Type::ArrayOf[String] # @!parse # # @return [Array] @@ -50,7 +50,7 @@ class VectorStoreCreateParams < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!attribute [r] name # The name of the vector store. @@ -82,9 +82,9 @@ class VectorStoreCreateParams < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class ExpiresAfter < OpenAI::BaseModel + class ExpiresAfter < OpenAI::Internal::Type::BaseModel # @!attribute anchor # Anchor timestamp after which the expiration policy applies. Supported anchors: # `last_active_at`. @@ -106,7 +106,7 @@ class ExpiresAfter < OpenAI::BaseModel # # # def initialize(days:, anchor: :last_active_at, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/vector_store_delete_params.rb b/lib/openai/models/vector_store_delete_params.rb index 42f0e18c..5beedec7 100644 --- a/lib/openai/models/vector_store_delete_params.rb +++ b/lib/openai/models/vector_store_delete_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::VectorStores#delete - class VectorStoreDeleteParams < OpenAI::BaseModel + class VectorStoreDeleteParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -13,7 +13,7 @@ class VectorStoreDeleteParams < OpenAI::BaseModel # # # def initialize(request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/vector_store_deleted.rb b/lib/openai/models/vector_store_deleted.rb index 84da9461..c3b6e2ab 100644 --- a/lib/openai/models/vector_store_deleted.rb +++ b/lib/openai/models/vector_store_deleted.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::VectorStores#delete - class VectorStoreDeleted < OpenAI::BaseModel + class VectorStoreDeleted < OpenAI::Internal::Type::BaseModel # @!attribute id # # @return [String] @@ -12,7 +12,7 @@ class VectorStoreDeleted < OpenAI::BaseModel # @!attribute deleted # # @return [Boolean] - required :deleted, OpenAI::BooleanModel + required :deleted, OpenAI::Internal::Type::BooleanModel # @!attribute object # @@ -26,7 +26,7 @@ class VectorStoreDeleted < OpenAI::BaseModel # # # def initialize(id:, deleted:, object: :"vector_store.deleted", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/vector_store_list_params.rb b/lib/openai/models/vector_store_list_params.rb index 12ea7117..eaea91f5 100644 --- a/lib/openai/models/vector_store_list_params.rb +++ b/lib/openai/models/vector_store_list_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::VectorStores#list - class VectorStoreListParams < OpenAI::BaseModel + class VectorStoreListParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -65,12 +65,12 @@ class VectorStoreListParams < OpenAI::BaseModel # # # def initialize(after: nil, before: nil, limit: nil, order: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Sort order by the `created_at` timestamp of the objects. `asc` for ascending # order and `desc` for descending order. module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASC = :asc DESC = :desc diff --git a/lib/openai/models/vector_store_retrieve_params.rb b/lib/openai/models/vector_store_retrieve_params.rb index 096ded4f..f3b3b098 100644 --- a/lib/openai/models/vector_store_retrieve_params.rb +++ b/lib/openai/models/vector_store_retrieve_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::VectorStores#retrieve - class VectorStoreRetrieveParams < OpenAI::BaseModel + class VectorStoreRetrieveParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -13,7 +13,7 @@ class VectorStoreRetrieveParams < OpenAI::BaseModel # # # def initialize(request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/vector_store_search_params.rb b/lib/openai/models/vector_store_search_params.rb index acf25855..e967edbe 100644 --- a/lib/openai/models/vector_store_search_params.rb +++ b/lib/openai/models/vector_store_search_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::VectorStores#search - class VectorStoreSearchParams < OpenAI::BaseModel + class VectorStoreSearchParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -49,7 +49,7 @@ class VectorStoreSearchParams < OpenAI::BaseModel # Whether to rewrite the natural language query for vector search. # # @return [Boolean, nil] - optional :rewrite_query, OpenAI::BooleanModel + optional :rewrite_query, OpenAI::Internal::Type::BooleanModel # @!parse # # @return [Boolean] @@ -75,11 +75,11 @@ class VectorStoreSearchParams < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # A query string for a search module Query - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String @@ -89,12 +89,12 @@ module Query # # @return [Array(String, Array)] # def self.variants; end - StringArray = OpenAI::ArrayOf[String] + StringArray = OpenAI::Internal::Type::ArrayOf[String] end # A filter to apply based on file attributes. module Filters - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # A filter used to compare a specified attribute key to a given value using a defined comparison operation. variant -> { OpenAI::Models::ComparisonFilter } @@ -107,7 +107,7 @@ module Filters # def self.variants; end end - class RankingOptions < OpenAI::BaseModel + class RankingOptions < OpenAI::Internal::Type::BaseModel # @!attribute [r] ranker # # @return [Symbol, OpenAI::Models::VectorStoreSearchParams::RankingOptions::Ranker, nil] @@ -134,11 +134,11 @@ class RankingOptions < OpenAI::BaseModel # # # def initialize(ranker: nil, score_threshold: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::VectorStoreSearchParams::RankingOptions#ranker module Ranker - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO = :auto DEFAULT_2024_11_15 = :"default-2024-11-15" diff --git a/lib/openai/models/vector_store_search_response.rb b/lib/openai/models/vector_store_search_response.rb index 71b724a2..785e82a6 100644 --- a/lib/openai/models/vector_store_search_response.rb +++ b/lib/openai/models/vector_store_search_response.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::VectorStores#search - class VectorStoreSearchResponse < OpenAI::BaseModel + class VectorStoreSearchResponse < OpenAI::Internal::Type::BaseModel # @!attribute attributes # Set of 16 key-value pairs that can be attached to an object. This can be useful # for storing additional information about the object in a structured format, and @@ -13,14 +13,15 @@ class VectorStoreSearchResponse < OpenAI::BaseModel # # @return [Hash{Symbol=>String, Float, Boolean}, nil] required :attributes, - -> { OpenAI::HashOf[union: OpenAI::Models::VectorStoreSearchResponse::Attribute] }, + -> { OpenAI::Internal::Type::HashOf[union: OpenAI::Models::VectorStoreSearchResponse::Attribute] }, nil?: true # @!attribute content # Content chunks from the file. # # @return [Array] - required :content, -> { OpenAI::ArrayOf[OpenAI::Models::VectorStoreSearchResponse::Content] } + required :content, + -> { OpenAI::Internal::Type::ArrayOf[OpenAI::Models::VectorStoreSearchResponse::Content] } # @!attribute file_id # The ID of the vector store file. @@ -49,23 +50,23 @@ class VectorStoreSearchResponse < OpenAI::BaseModel # # # def initialize(attributes:, content:, file_id:, filename:, score:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void module Attribute - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String variant Float - variant OpenAI::BooleanModel + variant OpenAI::Internal::Type::BooleanModel # @!parse # # @return [Array(String, Float, Boolean)] # def self.variants; end end - class Content < OpenAI::BaseModel + class Content < OpenAI::Internal::Type::BaseModel # @!attribute text # The text content returned from search. # @@ -84,13 +85,13 @@ class Content < OpenAI::BaseModel # # # def initialize(text:, type:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # The type of content. # # @see OpenAI::Models::VectorStoreSearchResponse::Content#type module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT = :text diff --git a/lib/openai/models/vector_store_update_params.rb b/lib/openai/models/vector_store_update_params.rb index 43edb6ab..46a09e25 100644 --- a/lib/openai/models/vector_store_update_params.rb +++ b/lib/openai/models/vector_store_update_params.rb @@ -3,7 +3,7 @@ module OpenAI module Models # @see OpenAI::Resources::VectorStores#update - class VectorStoreUpdateParams < OpenAI::BaseModel + class VectorStoreUpdateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -23,7 +23,7 @@ class VectorStoreUpdateParams < OpenAI::BaseModel # a maximum length of 512 characters. # # @return [Hash{Symbol=>String}, nil] - optional :metadata, OpenAI::HashOf[String], nil?: true + optional :metadata, OpenAI::Internal::Type::HashOf[String], nil?: true # @!attribute name # The name of the vector store. @@ -39,9 +39,9 @@ class VectorStoreUpdateParams < OpenAI::BaseModel # # # def initialize(expires_after: nil, metadata: nil, name: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void - class ExpiresAfter < OpenAI::BaseModel + class ExpiresAfter < OpenAI::Internal::Type::BaseModel # @!attribute anchor # Anchor timestamp after which the expiration policy applies. Supported anchors: # `last_active_at`. @@ -63,7 +63,7 @@ class ExpiresAfter < OpenAI::BaseModel # # # def initialize(days:, anchor: :last_active_at, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/vector_stores/file_batch_cancel_params.rb b/lib/openai/models/vector_stores/file_batch_cancel_params.rb index c87c2b85..a64a4b9c 100644 --- a/lib/openai/models/vector_stores/file_batch_cancel_params.rb +++ b/lib/openai/models/vector_stores/file_batch_cancel_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module VectorStores # @see OpenAI::Resources::VectorStores::FileBatches#cancel - class FileBatchCancelParams < OpenAI::BaseModel + class FileBatchCancelParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -20,7 +20,7 @@ class FileBatchCancelParams < OpenAI::BaseModel # # # def initialize(vector_store_id:, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/vector_stores/file_batch_create_params.rb b/lib/openai/models/vector_stores/file_batch_create_params.rb index 19b69c38..4140d9b9 100644 --- a/lib/openai/models/vector_stores/file_batch_create_params.rb +++ b/lib/openai/models/vector_stores/file_batch_create_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module VectorStores # @see OpenAI::Resources::VectorStores::FileBatches#create - class FileBatchCreateParams < OpenAI::BaseModel + class FileBatchCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -15,7 +15,7 @@ class FileBatchCreateParams < OpenAI::BaseModel # files. # # @return [Array] - required :file_ids, OpenAI::ArrayOf[String] + required :file_ids, OpenAI::Internal::Type::ArrayOf[String] # @!attribute attributes # Set of 16 key-value pairs that can be attached to an object. This can be useful @@ -26,7 +26,7 @@ class FileBatchCreateParams < OpenAI::BaseModel # # @return [Hash{Symbol=>String, Float, Boolean}, nil] optional :attributes, - -> { OpenAI::HashOf[union: OpenAI::Models::VectorStores::FileBatchCreateParams::Attribute] }, + -> { OpenAI::Internal::Type::HashOf[union: OpenAI::Models::VectorStores::FileBatchCreateParams::Attribute] }, nil?: true # @!attribute [r] chunking_strategy @@ -48,16 +48,16 @@ class FileBatchCreateParams < OpenAI::BaseModel # # # def initialize(file_ids:, attributes: nil, chunking_strategy: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void module Attribute - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String variant Float - variant OpenAI::BooleanModel + variant OpenAI::Internal::Type::BooleanModel # @!parse # # @return [Array(String, Float, Boolean)] diff --git a/lib/openai/models/vector_stores/file_batch_list_files_params.rb b/lib/openai/models/vector_stores/file_batch_list_files_params.rb index d19506f4..991ac9cc 100644 --- a/lib/openai/models/vector_stores/file_batch_list_files_params.rb +++ b/lib/openai/models/vector_stores/file_batch_list_files_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module VectorStores # @see OpenAI::Resources::VectorStores::FileBatches#list_files - class FileBatchListFilesParams < OpenAI::BaseModel + class FileBatchListFilesParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -83,11 +83,11 @@ class FileBatchListFilesParams < OpenAI::BaseModel # # # def initialize(vector_store_id:, after: nil, before: nil, filter: nil, limit: nil, order: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Filter by file status. One of `in_progress`, `completed`, `failed`, `cancelled`. module Filter - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS = :in_progress COMPLETED = :completed @@ -104,7 +104,7 @@ module Filter # Sort order by the `created_at` timestamp of the objects. `asc` for ascending # order and `desc` for descending order. module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASC = :asc DESC = :desc diff --git a/lib/openai/models/vector_stores/file_batch_retrieve_params.rb b/lib/openai/models/vector_stores/file_batch_retrieve_params.rb index a05da932..55499481 100644 --- a/lib/openai/models/vector_stores/file_batch_retrieve_params.rb +++ b/lib/openai/models/vector_stores/file_batch_retrieve_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module VectorStores # @see OpenAI::Resources::VectorStores::FileBatches#retrieve - class FileBatchRetrieveParams < OpenAI::BaseModel + class FileBatchRetrieveParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -20,7 +20,7 @@ class FileBatchRetrieveParams < OpenAI::BaseModel # # # def initialize(vector_store_id:, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/vector_stores/file_content_params.rb b/lib/openai/models/vector_stores/file_content_params.rb index 9b6fefcc..26be8b94 100644 --- a/lib/openai/models/vector_stores/file_content_params.rb +++ b/lib/openai/models/vector_stores/file_content_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module VectorStores # @see OpenAI::Resources::VectorStores::Files#content - class FileContentParams < OpenAI::BaseModel + class FileContentParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -20,7 +20,7 @@ class FileContentParams < OpenAI::BaseModel # # # def initialize(vector_store_id:, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/vector_stores/file_content_response.rb b/lib/openai/models/vector_stores/file_content_response.rb index 6f667726..01c19380 100644 --- a/lib/openai/models/vector_stores/file_content_response.rb +++ b/lib/openai/models/vector_stores/file_content_response.rb @@ -4,7 +4,7 @@ module OpenAI module Models module VectorStores # @see OpenAI::Resources::VectorStores::Files#content - class FileContentResponse < OpenAI::BaseModel + class FileContentResponse < OpenAI::Internal::Type::BaseModel # @!attribute [r] text # The text content # @@ -31,7 +31,7 @@ class FileContentResponse < OpenAI::BaseModel # # # def initialize(text: nil, type: nil, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/vector_stores/file_create_params.rb b/lib/openai/models/vector_stores/file_create_params.rb index 429f1ebc..f74f4ce2 100644 --- a/lib/openai/models/vector_stores/file_create_params.rb +++ b/lib/openai/models/vector_stores/file_create_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module VectorStores # @see OpenAI::Resources::VectorStores::Files#create - class FileCreateParams < OpenAI::BaseModel + class FileCreateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -26,7 +26,7 @@ class FileCreateParams < OpenAI::BaseModel # # @return [Hash{Symbol=>String, Float, Boolean}, nil] optional :attributes, - -> { OpenAI::HashOf[union: OpenAI::Models::VectorStores::FileCreateParams::Attribute] }, + -> { OpenAI::Internal::Type::HashOf[union: OpenAI::Models::VectorStores::FileCreateParams::Attribute] }, nil?: true # @!attribute [r] chunking_strategy @@ -48,16 +48,16 @@ class FileCreateParams < OpenAI::BaseModel # # # def initialize(file_id:, attributes: nil, chunking_strategy: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void module Attribute - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String variant Float - variant OpenAI::BooleanModel + variant OpenAI::Internal::Type::BooleanModel # @!parse # # @return [Array(String, Float, Boolean)] diff --git a/lib/openai/models/vector_stores/file_delete_params.rb b/lib/openai/models/vector_stores/file_delete_params.rb index bf6c7241..e4bee072 100644 --- a/lib/openai/models/vector_stores/file_delete_params.rb +++ b/lib/openai/models/vector_stores/file_delete_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module VectorStores # @see OpenAI::Resources::VectorStores::Files#delete - class FileDeleteParams < OpenAI::BaseModel + class FileDeleteParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -20,7 +20,7 @@ class FileDeleteParams < OpenAI::BaseModel # # # def initialize(vector_store_id:, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/vector_stores/file_list_params.rb b/lib/openai/models/vector_stores/file_list_params.rb index 01da7de3..0bcfb9d3 100644 --- a/lib/openai/models/vector_stores/file_list_params.rb +++ b/lib/openai/models/vector_stores/file_list_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module VectorStores # @see OpenAI::Resources::VectorStores::Files#list - class FileListParams < OpenAI::BaseModel + class FileListParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -77,11 +77,11 @@ class FileListParams < OpenAI::BaseModel # # # def initialize(after: nil, before: nil, filter: nil, limit: nil, order: nil, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # Filter by file status. One of `in_progress`, `completed`, `failed`, `cancelled`. module Filter - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS = :in_progress COMPLETED = :completed @@ -98,7 +98,7 @@ module Filter # Sort order by the `created_at` timestamp of the objects. `asc` for ascending # order and `desc` for descending order. module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASC = :asc DESC = :desc diff --git a/lib/openai/models/vector_stores/file_retrieve_params.rb b/lib/openai/models/vector_stores/file_retrieve_params.rb index 75f2cc40..9d9c26a7 100644 --- a/lib/openai/models/vector_stores/file_retrieve_params.rb +++ b/lib/openai/models/vector_stores/file_retrieve_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module VectorStores # @see OpenAI::Resources::VectorStores::Files#retrieve - class FileRetrieveParams < OpenAI::BaseModel + class FileRetrieveParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -20,7 +20,7 @@ class FileRetrieveParams < OpenAI::BaseModel # # # def initialize(vector_store_id:, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end end diff --git a/lib/openai/models/vector_stores/file_update_params.rb b/lib/openai/models/vector_stores/file_update_params.rb index a14267b5..0cea5737 100644 --- a/lib/openai/models/vector_stores/file_update_params.rb +++ b/lib/openai/models/vector_stores/file_update_params.rb @@ -4,7 +4,7 @@ module OpenAI module Models module VectorStores # @see OpenAI::Resources::VectorStores::Files#update - class FileUpdateParams < OpenAI::BaseModel + class FileUpdateParams < OpenAI::Internal::Type::BaseModel # @!parse # extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -23,7 +23,7 @@ class FileUpdateParams < OpenAI::BaseModel # # @return [Hash{Symbol=>String, Float, Boolean}, nil] required :attributes, - -> { OpenAI::HashOf[union: OpenAI::Models::VectorStores::FileUpdateParams::Attribute] }, + -> { OpenAI::Internal::Type::HashOf[union: OpenAI::Models::VectorStores::FileUpdateParams::Attribute] }, nil?: true # @!parse @@ -33,16 +33,16 @@ class FileUpdateParams < OpenAI::BaseModel # # # def initialize(vector_store_id:, attributes:, request_options: {}, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void module Attribute - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String variant Float - variant OpenAI::BooleanModel + variant OpenAI::Internal::Type::BooleanModel # @!parse # # @return [Array(String, Float, Boolean)] diff --git a/lib/openai/models/vector_stores/vector_store_file.rb b/lib/openai/models/vector_stores/vector_store_file.rb index 7c64487b..12ed8106 100644 --- a/lib/openai/models/vector_stores/vector_store_file.rb +++ b/lib/openai/models/vector_stores/vector_store_file.rb @@ -4,7 +4,7 @@ module OpenAI module Models module VectorStores # @see OpenAI::Resources::VectorStores::Files#create - class VectorStoreFile < OpenAI::BaseModel + class VectorStoreFile < OpenAI::Internal::Type::BaseModel # @!attribute id # The identifier, which can be referenced in API endpoints. # @@ -63,7 +63,7 @@ class VectorStoreFile < OpenAI::BaseModel # # @return [Hash{Symbol=>String, Float, Boolean}, nil] optional :attributes, - -> { OpenAI::HashOf[union: OpenAI::Models::VectorStores::VectorStoreFile::Attribute] }, + -> { OpenAI::Internal::Type::HashOf[union: OpenAI::Models::VectorStores::VectorStoreFile::Attribute] }, nil?: true # @!attribute [r] chunking_strategy @@ -104,10 +104,10 @@ class VectorStoreFile < OpenAI::BaseModel # super # end - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::VectorStores::VectorStoreFile#last_error - class LastError < OpenAI::BaseModel + class LastError < OpenAI::Internal::Type::BaseModel # @!attribute code # One of `server_error` or `rate_limit_exceeded`. # @@ -129,13 +129,13 @@ class LastError < OpenAI::BaseModel # # # def initialize(code:, message:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # One of `server_error` or `rate_limit_exceeded`. # # @see OpenAI::Models::VectorStores::VectorStoreFile::LastError#code module Code - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum SERVER_ERROR = :server_error UNSUPPORTED_FILE = :unsupported_file @@ -155,7 +155,7 @@ module Code # # @see OpenAI::Models::VectorStores::VectorStoreFile#status module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS = :in_progress COMPLETED = :completed @@ -170,13 +170,13 @@ module Status end module Attribute - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant String variant Float - variant OpenAI::BooleanModel + variant OpenAI::Internal::Type::BooleanModel # @!parse # # @return [Array(String, Float, Boolean)] diff --git a/lib/openai/models/vector_stores/vector_store_file_batch.rb b/lib/openai/models/vector_stores/vector_store_file_batch.rb index b08f672c..ebcd7390 100644 --- a/lib/openai/models/vector_stores/vector_store_file_batch.rb +++ b/lib/openai/models/vector_stores/vector_store_file_batch.rb @@ -4,7 +4,7 @@ module OpenAI module Models module VectorStores # @see OpenAI::Resources::VectorStores::FileBatches#create - class VectorStoreFileBatch < OpenAI::BaseModel + class VectorStoreFileBatch < OpenAI::Internal::Type::BaseModel # @!attribute id # The identifier, which can be referenced in API endpoints. # @@ -57,10 +57,10 @@ class VectorStoreFileBatch < OpenAI::BaseModel # # # def initialize(id:, created_at:, file_counts:, status:, vector_store_id:, object: :"vector_store.files_batch", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void # @see OpenAI::Models::VectorStores::VectorStoreFileBatch#file_counts - class FileCounts < OpenAI::BaseModel + class FileCounts < OpenAI::Internal::Type::BaseModel # @!attribute cancelled # The number of files that where cancelled. # @@ -100,7 +100,7 @@ class FileCounts < OpenAI::BaseModel # # # def initialize(cancelled:, completed:, failed:, in_progress:, total:, **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end # The status of the vector store files batch, which can be either `in_progress`, @@ -108,7 +108,7 @@ class FileCounts < OpenAI::BaseModel # # @see OpenAI::Models::VectorStores::VectorStoreFileBatch#status module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS = :in_progress COMPLETED = :completed diff --git a/lib/openai/models/vector_stores/vector_store_file_deleted.rb b/lib/openai/models/vector_stores/vector_store_file_deleted.rb index 26fa7214..13f5c954 100644 --- a/lib/openai/models/vector_stores/vector_store_file_deleted.rb +++ b/lib/openai/models/vector_stores/vector_store_file_deleted.rb @@ -4,7 +4,7 @@ module OpenAI module Models module VectorStores # @see OpenAI::Resources::VectorStores::Files#delete - class VectorStoreFileDeleted < OpenAI::BaseModel + class VectorStoreFileDeleted < OpenAI::Internal::Type::BaseModel # @!attribute id # # @return [String] @@ -13,7 +13,7 @@ class VectorStoreFileDeleted < OpenAI::BaseModel # @!attribute deleted # # @return [Boolean] - required :deleted, OpenAI::BooleanModel + required :deleted, OpenAI::Internal::Type::BooleanModel # @!attribute object # @@ -27,7 +27,7 @@ class VectorStoreFileDeleted < OpenAI::BaseModel # # # def initialize(id:, deleted:, object: :"vector_store.file.deleted", **) = super - # def initialize: (Hash | OpenAI::BaseModel) -> void + # def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void end end diff --git a/lib/openai/request_options.rb b/lib/openai/request_options.rb index 279c2936..5f374fb2 100644 --- a/lib/openai/request_options.rb +++ b/lib/openai/request_options.rb @@ -6,7 +6,7 @@ module OpenAI # # When making a request, you can pass an actual {RequestOptions} instance, or # simply pass a Hash with symbol keys matching the attributes on this class. - class RequestOptions < OpenAI::BaseModel + class RequestOptions < OpenAI::Internal::Type::BaseModel # @api private # # @param opts [OpenAI::RequestOptions, Hash{Symbol=>Object}] @@ -37,21 +37,21 @@ def self.validate!(opts) # `query` given at the client level. # # @return [Hash{String=>Array, String, nil}, nil] - optional :extra_query, OpenAI::HashOf[OpenAI::ArrayOf[String]] + optional :extra_query, OpenAI::Internal::Type::HashOf[OpenAI::Internal::Type::ArrayOf[String]] # @!attribute extra_headers # Extra headers to send with the request. These are `.merged`’d into any # `extra_headers` given at the client level. # # @return [Hash{String=>String, nil}, nil] - optional :extra_headers, OpenAI::HashOf[String, nil?: true] + optional :extra_headers, OpenAI::Internal::Type::HashOf[String, nil?: true] # @!attribute extra_body # Extra data to send with the request. These are deep merged into any data # generated as part of the normal request. # # @return [Object, nil] - optional :extra_body, OpenAI::HashOf[OpenAI::Unknown] + optional :extra_body, OpenAI::Internal::Type::HashOf[OpenAI::Internal::Type::Unknown] # @!attribute max_retries # Maximum number of retries to attempt after a failed initial request. diff --git a/rbi/lib/openai/aliases.rbi b/rbi/lib/openai/aliases.rbi deleted file mode 100644 index 5641bd39..00000000 --- a/rbi/lib/openai/aliases.rbi +++ /dev/null @@ -1,19 +0,0 @@ -# typed: strong - -module OpenAI - Unknown = OpenAI::Internal::Type::Unknown - - BooleanModel = OpenAI::Internal::Type::BooleanModel - - Enum = OpenAI::Internal::Type::Enum - - Union = OpenAI::Internal::Type::Union - - ArrayOf = OpenAI::Internal::Type::ArrayOf - - HashOf = OpenAI::Internal::Type::HashOf - - BaseModel = OpenAI::Internal::Type::BaseModel - - RequestParameters = OpenAI::Internal::Type::RequestParameters -end diff --git a/rbi/lib/openai/errors.rbi b/rbi/lib/openai/errors.rbi index d04fc102..a0f3280e 100644 --- a/rbi/lib/openai/errors.rbi +++ b/rbi/lib/openai/errors.rbi @@ -170,32 +170,4 @@ module OpenAI HTTP_STATUS = T.let((500..), T::Range[Integer]) end end - - Error = OpenAI::Errors::Error - - ConversionError = OpenAI::Errors::ConversionError - - APIError = OpenAI::Errors::APIError - - APIStatusError = OpenAI::Errors::APIStatusError - - APIConnectionError = OpenAI::Errors::APIConnectionError - - APITimeoutError = OpenAI::Errors::APITimeoutError - - BadRequestError = OpenAI::Errors::BadRequestError - - AuthenticationError = OpenAI::Errors::AuthenticationError - - PermissionDeniedError = OpenAI::Errors::PermissionDeniedError - - NotFoundError = OpenAI::Errors::NotFoundError - - ConflictError = OpenAI::Errors::ConflictError - - UnprocessableEntityError = OpenAI::Errors::UnprocessableEntityError - - RateLimitError = OpenAI::Errors::RateLimitError - - InternalServerError = OpenAI::Errors::InternalServerError end diff --git a/rbi/lib/openai/internal.rbi b/rbi/lib/openai/internal.rbi new file mode 100644 index 00000000..ff17c5f5 --- /dev/null +++ b/rbi/lib/openai/internal.rbi @@ -0,0 +1,10 @@ +# typed: strong + +module OpenAI + # @api private + module Internal + # Due to the current WIP status of Shapes support in Sorbet, types referencing + # this alias might be refined in the future. + AnyHash = T.type_alias { T::Hash[Symbol, T.anything] } + end +end diff --git a/rbi/lib/openai/internal/transport/base_client.rbi b/rbi/lib/openai/internal/transport/base_client.rbi index abc94734..f27429e0 100644 --- a/rbi/lib/openai/internal/transport/base_client.rbi +++ b/rbi/lib/openai/internal/transport/base_client.rbi @@ -25,10 +25,13 @@ module OpenAI ), body: T.nilable(T.anything), unwrap: T.nilable(Symbol), - page: T.nilable(T::Class[OpenAI::Internal::Type::BasePage[OpenAI::BaseModel]]), - stream: T.nilable(T::Class[OpenAI::Internal::Type::BaseStream[T.anything, OpenAI::BaseModel]]), + page: T.nilable(T::Class[OpenAI::Internal::Type::BasePage[OpenAI::Internal::Type::BaseModel]]), + stream: T.nilable( + T::Class[OpenAI::Internal::Type::BaseStream[T.anything, + OpenAI::Internal::Type::BaseModel]] + ), model: T.nilable(OpenAI::Internal::Type::Converter::Input), - options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) } end @@ -136,7 +139,7 @@ module OpenAI overridable .params( req: OpenAI::Internal::Transport::BaseClient::RequestComponentsShape, - opts: OpenAI::Internal::Util::AnyHash + opts: OpenAI::Internal::AnyHash ) .returns(OpenAI::Internal::Transport::BaseClient::RequestInputShape) end @@ -180,10 +183,13 @@ module OpenAI ), body: T.nilable(T.anything), unwrap: T.nilable(Symbol), - page: T.nilable(T::Class[OpenAI::Internal::Type::BasePage[OpenAI::BaseModel]]), - stream: T.nilable(T::Class[OpenAI::Internal::Type::BaseStream[T.anything, OpenAI::BaseModel]]), + page: T.nilable(T::Class[OpenAI::Internal::Type::BasePage[OpenAI::Internal::Type::BaseModel]]), + stream: T.nilable( + T::Class[OpenAI::Internal::Type::BaseStream[T.anything, + OpenAI::Internal::Type::BaseModel]] + ), model: T.nilable(OpenAI::Internal::Type::Converter::Input), - options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(T.anything) end @@ -196,7 +202,7 @@ module OpenAI unwrap: nil, page: nil, stream: nil, - model: OpenAI::Unknown, + model: OpenAI::Internal::Type::Unknown, options: {} ) end diff --git a/rbi/lib/openai/internal/type/array_of.rbi b/rbi/lib/openai/internal/type/array_of.rbi index 54f032b7..d01b1212 100644 --- a/rbi/lib/openai/internal/type/array_of.rbi +++ b/rbi/lib/openai/internal/type/array_of.rbi @@ -17,11 +17,11 @@ module OpenAI sig(:final) do params( type_info: T.any( - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, T.proc.returns(OpenAI::Internal::Type::Converter::Input), OpenAI::Internal::Type::Converter::Input ), - spec: OpenAI::Internal::Util::AnyHash + spec: OpenAI::Internal::AnyHash ) .returns(T.attached_class) end @@ -72,11 +72,11 @@ module OpenAI sig(:final) do params( type_info: T.any( - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, T.proc.returns(OpenAI::Internal::Type::Converter::Input), OpenAI::Internal::Type::Converter::Input ), - spec: OpenAI::Internal::Util::AnyHash + spec: OpenAI::Internal::AnyHash ) .void end diff --git a/rbi/lib/openai/internal/type/base_model.rbi b/rbi/lib/openai/internal/type/base_model.rbi index 6a8646a9..f5cf4f9c 100644 --- a/rbi/lib/openai/internal/type/base_model.rbi +++ b/rbi/lib/openai/internal/type/base_model.rbi @@ -22,7 +22,7 @@ module OpenAI T::Hash[ Symbol, T.all( - OpenAI::BaseModel::KnownFieldShape, + OpenAI::Internal::Type::BaseModel::KnownFieldShape, {type_fn: T.proc.returns(OpenAI::Internal::Type::Converter::Input)} ) ] @@ -34,11 +34,13 @@ module OpenAI # @api private sig do returns( - T::Hash[Symbol, - T.all( - OpenAI::BaseModel::KnownFieldShape, - {type: OpenAI::Internal::Type::Converter::Input} - )] + T::Hash[ + Symbol, + T.all( + OpenAI::Internal::Type::BaseModel::KnownFieldShape, + {type: OpenAI::Internal::Type::Converter::Input} + ) + ] ) end def fields @@ -60,7 +62,7 @@ module OpenAI T.proc.returns(OpenAI::Internal::Type::Converter::Input), OpenAI::Internal::Type::Converter::Input ), - spec: OpenAI::Internal::Util::AnyHash + spec: OpenAI::Internal::AnyHash ) .void end @@ -72,11 +74,11 @@ module OpenAI params( name_sym: Symbol, type_info: T.any( - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, T.proc.returns(OpenAI::Internal::Type::Converter::Input), OpenAI::Internal::Type::Converter::Input ), - spec: OpenAI::Internal::Util::AnyHash + spec: OpenAI::Internal::AnyHash ) .void end @@ -88,11 +90,11 @@ module OpenAI params( name_sym: Symbol, type_info: T.any( - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, T.proc.returns(OpenAI::Internal::Type::Converter::Input), OpenAI::Internal::Type::Converter::Input ), - spec: OpenAI::Internal::Util::AnyHash + spec: OpenAI::Internal::AnyHash ) .void end @@ -128,7 +130,7 @@ module OpenAI sig do override .params( - value: T.any(OpenAI::BaseModel, T::Hash[T.anything, T.anything], T.anything), + value: T.any(OpenAI::Internal::Type::BaseModel, T::Hash[T.anything, T.anything], T.anything), state: OpenAI::Internal::Type::Converter::State ) .returns(T.any(T.attached_class, T.anything)) @@ -164,7 +166,7 @@ module OpenAI # # This method is not recursive. The returned value is shared by the object, so it # should not be mutated. - sig { overridable.returns(OpenAI::Internal::Util::AnyHash) } + sig { overridable.returns(OpenAI::Internal::AnyHash) } def to_h end @@ -176,11 +178,11 @@ module OpenAI # # This method is not recursive. The returned value is shared by the object, so it # should not be mutated. - sig { overridable.returns(OpenAI::Internal::Util::AnyHash) } + sig { overridable.returns(OpenAI::Internal::AnyHash) } def to_hash end - sig { params(keys: T.nilable(T::Array[Symbol])).returns(OpenAI::Internal::Util::AnyHash) } + sig { params(keys: T.nilable(T::Array[Symbol])).returns(OpenAI::Internal::AnyHash) } def deconstruct_keys(keys) end diff --git a/rbi/lib/openai/internal/type/hash_of.rbi b/rbi/lib/openai/internal/type/hash_of.rbi index 207ff5c9..d63a83f4 100644 --- a/rbi/lib/openai/internal/type/hash_of.rbi +++ b/rbi/lib/openai/internal/type/hash_of.rbi @@ -17,11 +17,11 @@ module OpenAI sig(:final) do params( type_info: T.any( - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, T.proc.returns(OpenAI::Internal::Type::Converter::Input), OpenAI::Internal::Type::Converter::Input ), - spec: OpenAI::Internal::Util::AnyHash + spec: OpenAI::Internal::AnyHash ) .returns(T.attached_class) end @@ -43,7 +43,7 @@ module OpenAI value: T.any(T::Hash[T.anything, T.anything], T.anything), state: OpenAI::Internal::Type::Converter::State ) - .returns(T.any(OpenAI::Internal::Util::AnyHash, T.anything)) + .returns(T.any(OpenAI::Internal::AnyHash, T.anything)) end def coerce(value, state:) end @@ -52,7 +52,7 @@ module OpenAI sig(:final) do override .params(value: T.any(T::Hash[T.anything, T.anything], T.anything)) - .returns(T.any(OpenAI::Internal::Util::AnyHash, T.anything)) + .returns(T.any(OpenAI::Internal::AnyHash, T.anything)) end def dump(value) end @@ -71,11 +71,11 @@ module OpenAI sig(:final) do params( type_info: T.any( - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, T.proc.returns(OpenAI::Internal::Type::Converter::Input), OpenAI::Internal::Type::Converter::Input ), - spec: OpenAI::Internal::Util::AnyHash + spec: OpenAI::Internal::AnyHash ) .void end diff --git a/rbi/lib/openai/internal/type/request_parameters.rbi b/rbi/lib/openai/internal/type/request_parameters.rbi index 6170a658..18a79320 100644 --- a/rbi/lib/openai/internal/type/request_parameters.rbi +++ b/rbi/lib/openai/internal/type/request_parameters.rbi @@ -6,13 +6,13 @@ module OpenAI # @api private module RequestParameters # Options to specify HTTP behaviour for this request. - sig { returns(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) } + sig { returns(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) } attr_accessor :request_options # @api private module Converter # @api private - sig { params(params: T.anything).returns([T.anything, OpenAI::Internal::Util::AnyHash]) } + sig { params(params: T.anything).returns([T.anything, OpenAI::Internal::AnyHash]) } def dump_request(params) end end diff --git a/rbi/lib/openai/internal/type/union.rbi b/rbi/lib/openai/internal/type/union.rbi index d96795ec..3d1486f0 100644 --- a/rbi/lib/openai/internal/type/union.rbi +++ b/rbi/lib/openai/internal/type/union.rbi @@ -34,8 +34,8 @@ module OpenAI # @api private sig do params( - key: T.any(Symbol, OpenAI::Internal::Util::AnyHash, T.proc.returns(T.anything), T.anything), - spec: T.any(OpenAI::Internal::Util::AnyHash, T.proc.returns(T.anything), T.anything) + key: T.any(Symbol, OpenAI::Internal::AnyHash, T.proc.returns(T.anything), T.anything), + spec: T.any(OpenAI::Internal::AnyHash, T.proc.returns(T.anything), T.anything) ) .void end diff --git a/rbi/lib/openai/internal/util.rbi b/rbi/lib/openai/internal/util.rbi index 5414927e..7abcdc1a 100644 --- a/rbi/lib/openai/internal/util.rbi +++ b/rbi/lib/openai/internal/util.rbi @@ -4,10 +4,6 @@ module OpenAI module Internal # @api private module Util - # Due to the current WIP status of Shapes support in Sorbet, types referencing - # this alias might be refined in the future. - AnyHash = T.type_alias { T::Hash[Symbol, T.anything] } - # @api private sig { returns(Float) } def self.monotonic_secs @@ -57,13 +53,6 @@ module OpenAI end end - # Use this to indicate that a value should be explicitly removed from a data - # structure when using `OpenAI::Internal::Util.deep_merge`. - # - # e.g. merging `{a: 1}` and `{a: OMIT}` should produce `{}`, where merging - # `{a: 1}` and `{}` would produce `{a: 1}`. - OMIT = T.let(T.anything, T.anything) - class << self # @api private sig { params(lhs: T.anything, rhs: T.anything, concat: T::Boolean).returns(T.anything) } @@ -90,7 +79,7 @@ module OpenAI # @api private sig do params( - data: T.any(OpenAI::Internal::Util::AnyHash, T::Array[T.anything], T.anything), + data: T.any(OpenAI::Internal::AnyHash, T::Array[T.anything], T.anything), pick: T.nilable(T.any(Symbol, Integer, T::Array[T.any(Symbol, Integer)])), sentinel: T.nilable(T.anything), blk: T.nilable(T.proc.returns(T.anything)) diff --git a/rbi/lib/openai/models/all_models.rbi b/rbi/lib/openai/models/all_models.rbi index 3d16cbd5..edc0d8c5 100644 --- a/rbi/lib/openai/models/all_models.rbi +++ b/rbi/lib/openai/models/all_models.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module AllModels - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, OpenAI::Models::ChatModel::OrSymbol, OpenAI::Models::AllModels::OrSymbol]) } def self.variants diff --git a/rbi/lib/openai/models/audio/speech_create_params.rbi b/rbi/lib/openai/models/audio/speech_create_params.rbi index 4f531df7..08cd04a2 100644 --- a/rbi/lib/openai/models/audio/speech_create_params.rbi +++ b/rbi/lib/openai/models/audio/speech_create_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Audio - class SpeechCreateParams < OpenAI::BaseModel + class SpeechCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -55,7 +55,7 @@ module OpenAI instructions: String, response_format: OpenAI::Models::Audio::SpeechCreateParams::ResponseFormat::OrSymbol, speed: Float, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -90,7 +90,7 @@ module OpenAI # One of the available [TTS models](https://platform.openai.com/docs/models#tts): # `tts-1`, `tts-1-hd` or `gpt-4o-mini-tts`. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, OpenAI::Models::Audio::SpeechModel::OrSymbol]) } def self.variants @@ -102,7 +102,7 @@ module OpenAI # `verse`. Previews of the voices are available in the # [Text to speech guide](https://platform.openai.com/docs/guides/text-to-speech#voice-options). module Voice - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, OpenAI::Models::Audio::SpeechCreateParams::Voice::OrSymbol]) } def self.variants @@ -128,7 +128,7 @@ module OpenAI # The format to audio in. Supported formats are `mp3`, `opus`, `aac`, `flac`, # `wav`, and `pcm`. module ResponseFormat - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Audio::SpeechCreateParams::ResponseFormat) } OrSymbol = diff --git a/rbi/lib/openai/models/audio/speech_model.rbi b/rbi/lib/openai/models/audio/speech_model.rbi index ed1acea8..a64b1353 100644 --- a/rbi/lib/openai/models/audio/speech_model.rbi +++ b/rbi/lib/openai/models/audio/speech_model.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Audio module SpeechModel - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Audio::SpeechModel) } OrSymbol = T.type_alias { T.any(Symbol, String, OpenAI::Models::Audio::SpeechModel::TaggedSymbol) } diff --git a/rbi/lib/openai/models/audio/transcription.rbi b/rbi/lib/openai/models/audio/transcription.rbi index f6d6ac1a..5310fa78 100644 --- a/rbi/lib/openai/models/audio/transcription.rbi +++ b/rbi/lib/openai/models/audio/transcription.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Audio - class Transcription < OpenAI::BaseModel + class Transcription < OpenAI::Internal::Type::BaseModel # The transcribed text. sig { returns(String) } attr_accessor :text @@ -16,7 +16,7 @@ module OpenAI sig do params( - logprobs: T::Array[T.any(OpenAI::Models::Audio::Transcription::Logprob, OpenAI::Internal::Util::AnyHash)] + logprobs: T::Array[T.any(OpenAI::Models::Audio::Transcription::Logprob, OpenAI::Internal::AnyHash)] ) .void end @@ -27,7 +27,7 @@ module OpenAI sig do params( text: String, - logprobs: T::Array[T.any(OpenAI::Models::Audio::Transcription::Logprob, OpenAI::Internal::Util::AnyHash)] + logprobs: T::Array[T.any(OpenAI::Models::Audio::Transcription::Logprob, OpenAI::Internal::AnyHash)] ) .returns(T.attached_class) end @@ -38,7 +38,7 @@ module OpenAI def to_hash end - class Logprob < OpenAI::BaseModel + class Logprob < OpenAI::Internal::Type::BaseModel # The token in the transcription. sig { returns(T.nilable(String)) } attr_reader :token diff --git a/rbi/lib/openai/models/audio/transcription_create_params.rbi b/rbi/lib/openai/models/audio/transcription_create_params.rbi index 2db27d93..951cde7b 100644 --- a/rbi/lib/openai/models/audio/transcription_create_params.rbi +++ b/rbi/lib/openai/models/audio/transcription_create_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Audio - class TranscriptionCreateParams < OpenAI::BaseModel + class TranscriptionCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -98,7 +98,7 @@ module OpenAI response_format: OpenAI::Models::AudioResponseFormat::OrSymbol, temperature: Float, timestamp_granularities: T::Array[OpenAI::Models::Audio::TranscriptionCreateParams::TimestampGranularity::OrSymbol], - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -138,7 +138,7 @@ module OpenAI # `gpt-4o-mini-transcribe`, and `whisper-1` (which is powered by our open source # Whisper V2 model). module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, OpenAI::Models::AudioModel::OrSymbol]) } def self.variants @@ -146,7 +146,7 @@ module OpenAI end module TimestampGranularity - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Audio::TranscriptionCreateParams::TimestampGranularity) } diff --git a/rbi/lib/openai/models/audio/transcription_create_response.rbi b/rbi/lib/openai/models/audio/transcription_create_response.rbi index 482495e5..c2a3aad6 100644 --- a/rbi/lib/openai/models/audio/transcription_create_response.rbi +++ b/rbi/lib/openai/models/audio/transcription_create_response.rbi @@ -6,7 +6,7 @@ module OpenAI # Represents a transcription response returned by model, based on the provided # input. module TranscriptionCreateResponse - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([OpenAI::Models::Audio::Transcription, OpenAI::Models::Audio::TranscriptionVerbose]) } def self.variants diff --git a/rbi/lib/openai/models/audio/transcription_include.rbi b/rbi/lib/openai/models/audio/transcription_include.rbi index 878b37d7..e1de4abe 100644 --- a/rbi/lib/openai/models/audio/transcription_include.rbi +++ b/rbi/lib/openai/models/audio/transcription_include.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Audio module TranscriptionInclude - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Audio::TranscriptionInclude) } OrSymbol = diff --git a/rbi/lib/openai/models/audio/transcription_segment.rbi b/rbi/lib/openai/models/audio/transcription_segment.rbi index 96cdfd4e..b4b718a4 100644 --- a/rbi/lib/openai/models/audio/transcription_segment.rbi +++ b/rbi/lib/openai/models/audio/transcription_segment.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Audio - class TranscriptionSegment < OpenAI::BaseModel + class TranscriptionSegment < OpenAI::Internal::Type::BaseModel # Unique identifier of the segment. sig { returns(Integer) } attr_accessor :id diff --git a/rbi/lib/openai/models/audio/transcription_stream_event.rbi b/rbi/lib/openai/models/audio/transcription_stream_event.rbi index 82ef83d4..32426872 100644 --- a/rbi/lib/openai/models/audio/transcription_stream_event.rbi +++ b/rbi/lib/openai/models/audio/transcription_stream_event.rbi @@ -8,7 +8,7 @@ module OpenAI # [create a transcription](https://platform.openai.com/docs/api-reference/audio/create-transcription) # with the `Stream` parameter set to `true`. module TranscriptionStreamEvent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/audio/transcription_text_delta_event.rbi b/rbi/lib/openai/models/audio/transcription_text_delta_event.rbi index 7612edef..aeddfb85 100644 --- a/rbi/lib/openai/models/audio/transcription_text_delta_event.rbi +++ b/rbi/lib/openai/models/audio/transcription_text_delta_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Audio - class TranscriptionTextDeltaEvent < OpenAI::BaseModel + class TranscriptionTextDeltaEvent < OpenAI::Internal::Type::BaseModel # The text delta that was additionally transcribed. sig { returns(String) } attr_accessor :delta @@ -20,7 +20,7 @@ module OpenAI sig do params( - logprobs: T::Array[T.any(OpenAI::Models::Audio::TranscriptionTextDeltaEvent::Logprob, OpenAI::Internal::Util::AnyHash)] + logprobs: T::Array[T.any(OpenAI::Models::Audio::TranscriptionTextDeltaEvent::Logprob, OpenAI::Internal::AnyHash)] ) .void end @@ -33,7 +33,7 @@ module OpenAI sig do params( delta: String, - logprobs: T::Array[T.any(OpenAI::Models::Audio::TranscriptionTextDeltaEvent::Logprob, OpenAI::Internal::Util::AnyHash)], + logprobs: T::Array[T.any(OpenAI::Models::Audio::TranscriptionTextDeltaEvent::Logprob, OpenAI::Internal::AnyHash)], type: Symbol ) .returns(T.attached_class) @@ -50,7 +50,7 @@ module OpenAI def to_hash end - class Logprob < OpenAI::BaseModel + class Logprob < OpenAI::Internal::Type::BaseModel # The token that was used to generate the log probability. sig { returns(T.nilable(String)) } attr_reader :token diff --git a/rbi/lib/openai/models/audio/transcription_text_done_event.rbi b/rbi/lib/openai/models/audio/transcription_text_done_event.rbi index fd8ce791..c422dcf3 100644 --- a/rbi/lib/openai/models/audio/transcription_text_done_event.rbi +++ b/rbi/lib/openai/models/audio/transcription_text_done_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Audio - class TranscriptionTextDoneEvent < OpenAI::BaseModel + class TranscriptionTextDoneEvent < OpenAI::Internal::Type::BaseModel # The text that was transcribed. sig { returns(String) } attr_accessor :text @@ -21,7 +21,7 @@ module OpenAI sig do params( - logprobs: T::Array[T.any(OpenAI::Models::Audio::TranscriptionTextDoneEvent::Logprob, OpenAI::Internal::Util::AnyHash)] + logprobs: T::Array[T.any(OpenAI::Models::Audio::TranscriptionTextDoneEvent::Logprob, OpenAI::Internal::AnyHash)] ) .void end @@ -34,7 +34,7 @@ module OpenAI sig do params( text: String, - logprobs: T::Array[T.any(OpenAI::Models::Audio::TranscriptionTextDoneEvent::Logprob, OpenAI::Internal::Util::AnyHash)], + logprobs: T::Array[T.any(OpenAI::Models::Audio::TranscriptionTextDoneEvent::Logprob, OpenAI::Internal::AnyHash)], type: Symbol ) .returns(T.attached_class) @@ -51,7 +51,7 @@ module OpenAI def to_hash end - class Logprob < OpenAI::BaseModel + class Logprob < OpenAI::Internal::Type::BaseModel # The token that was used to generate the log probability. sig { returns(T.nilable(String)) } attr_reader :token diff --git a/rbi/lib/openai/models/audio/transcription_verbose.rbi b/rbi/lib/openai/models/audio/transcription_verbose.rbi index 6500d39d..cde99b0d 100644 --- a/rbi/lib/openai/models/audio/transcription_verbose.rbi +++ b/rbi/lib/openai/models/audio/transcription_verbose.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Audio - class TranscriptionVerbose < OpenAI::BaseModel + class TranscriptionVerbose < OpenAI::Internal::Type::BaseModel # The duration of the input audio. sig { returns(Float) } attr_accessor :duration @@ -21,9 +21,7 @@ module OpenAI attr_reader :segments sig do - params( - segments: T::Array[T.any(OpenAI::Models::Audio::TranscriptionSegment, OpenAI::Internal::Util::AnyHash)] - ) + params(segments: T::Array[T.any(OpenAI::Models::Audio::TranscriptionSegment, OpenAI::Internal::AnyHash)]) .void end attr_writer :segments @@ -32,10 +30,7 @@ module OpenAI sig { returns(T.nilable(T::Array[OpenAI::Models::Audio::TranscriptionWord])) } attr_reader :words - sig do - params(words: T::Array[T.any(OpenAI::Models::Audio::TranscriptionWord, OpenAI::Internal::Util::AnyHash)]) - .void - end + sig { params(words: T::Array[T.any(OpenAI::Models::Audio::TranscriptionWord, OpenAI::Internal::AnyHash)]).void } attr_writer :words # Represents a verbose json transcription response returned by model, based on the @@ -45,8 +40,8 @@ module OpenAI duration: Float, language: String, text: String, - segments: T::Array[T.any(OpenAI::Models::Audio::TranscriptionSegment, OpenAI::Internal::Util::AnyHash)], - words: T::Array[T.any(OpenAI::Models::Audio::TranscriptionWord, OpenAI::Internal::Util::AnyHash)] + segments: T::Array[T.any(OpenAI::Models::Audio::TranscriptionSegment, OpenAI::Internal::AnyHash)], + words: T::Array[T.any(OpenAI::Models::Audio::TranscriptionWord, OpenAI::Internal::AnyHash)] ) .returns(T.attached_class) end diff --git a/rbi/lib/openai/models/audio/transcription_word.rbi b/rbi/lib/openai/models/audio/transcription_word.rbi index c6adc3b1..5907efa6 100644 --- a/rbi/lib/openai/models/audio/transcription_word.rbi +++ b/rbi/lib/openai/models/audio/transcription_word.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Audio - class TranscriptionWord < OpenAI::BaseModel + class TranscriptionWord < OpenAI::Internal::Type::BaseModel # End time of the word in seconds. sig { returns(Float) } attr_accessor :end_ diff --git a/rbi/lib/openai/models/audio/translation.rbi b/rbi/lib/openai/models/audio/translation.rbi index 07767362..a2735384 100644 --- a/rbi/lib/openai/models/audio/translation.rbi +++ b/rbi/lib/openai/models/audio/translation.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Audio - class Translation < OpenAI::BaseModel + class Translation < OpenAI::Internal::Type::BaseModel sig { returns(String) } attr_accessor :text diff --git a/rbi/lib/openai/models/audio/translation_create_params.rbi b/rbi/lib/openai/models/audio/translation_create_params.rbi index 6c2cdb3a..aa1ac9b3 100644 --- a/rbi/lib/openai/models/audio/translation_create_params.rbi +++ b/rbi/lib/openai/models/audio/translation_create_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Audio - class TranslationCreateParams < OpenAI::BaseModel + class TranslationCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -53,7 +53,7 @@ module OpenAI prompt: String, response_format: OpenAI::Models::Audio::TranslationCreateParams::ResponseFormat::OrSymbol, temperature: Float, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -79,7 +79,7 @@ module OpenAI # ID of the model to use. Only `whisper-1` (which is powered by our open source # Whisper V2 model) is currently available. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, OpenAI::Models::AudioModel::OrSymbol]) } def self.variants @@ -89,7 +89,7 @@ module OpenAI # The format of the output, in one of these options: `json`, `text`, `srt`, # `verbose_json`, or `vtt`. module ResponseFormat - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Audio::TranslationCreateParams::ResponseFormat) } diff --git a/rbi/lib/openai/models/audio/translation_create_response.rbi b/rbi/lib/openai/models/audio/translation_create_response.rbi index 8c9c9851..8c96ec6e 100644 --- a/rbi/lib/openai/models/audio/translation_create_response.rbi +++ b/rbi/lib/openai/models/audio/translation_create_response.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Audio module TranslationCreateResponse - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([OpenAI::Models::Audio::Translation, OpenAI::Models::Audio::TranslationVerbose]) } def self.variants diff --git a/rbi/lib/openai/models/audio/translation_verbose.rbi b/rbi/lib/openai/models/audio/translation_verbose.rbi index fee32692..90251977 100644 --- a/rbi/lib/openai/models/audio/translation_verbose.rbi +++ b/rbi/lib/openai/models/audio/translation_verbose.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Audio - class TranslationVerbose < OpenAI::BaseModel + class TranslationVerbose < OpenAI::Internal::Type::BaseModel # The duration of the input audio. sig { returns(Float) } attr_accessor :duration @@ -21,9 +21,7 @@ module OpenAI attr_reader :segments sig do - params( - segments: T::Array[T.any(OpenAI::Models::Audio::TranscriptionSegment, OpenAI::Internal::Util::AnyHash)] - ) + params(segments: T::Array[T.any(OpenAI::Models::Audio::TranscriptionSegment, OpenAI::Internal::AnyHash)]) .void end attr_writer :segments @@ -33,7 +31,7 @@ module OpenAI duration: Float, language: String, text: String, - segments: T::Array[T.any(OpenAI::Models::Audio::TranscriptionSegment, OpenAI::Internal::Util::AnyHash)] + segments: T::Array[T.any(OpenAI::Models::Audio::TranscriptionSegment, OpenAI::Internal::AnyHash)] ) .returns(T.attached_class) end diff --git a/rbi/lib/openai/models/audio_model.rbi b/rbi/lib/openai/models/audio_model.rbi index 656d0083..09edaa97 100644 --- a/rbi/lib/openai/models/audio_model.rbi +++ b/rbi/lib/openai/models/audio_model.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module AudioModel - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::AudioModel) } OrSymbol = T.type_alias { T.any(Symbol, String, OpenAI::Models::AudioModel::TaggedSymbol) } diff --git a/rbi/lib/openai/models/audio_response_format.rbi b/rbi/lib/openai/models/audio_response_format.rbi index 1246abca..ac5e8f58 100644 --- a/rbi/lib/openai/models/audio_response_format.rbi +++ b/rbi/lib/openai/models/audio_response_format.rbi @@ -6,7 +6,7 @@ module OpenAI # `verbose_json`, or `vtt`. For `gpt-4o-transcribe` and `gpt-4o-mini-transcribe`, # the only supported format is `json`. module AudioResponseFormat - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::AudioResponseFormat) } OrSymbol = T.type_alias { T.any(Symbol, String, OpenAI::Models::AudioResponseFormat::TaggedSymbol) } diff --git a/rbi/lib/openai/models/auto_file_chunking_strategy_param.rbi b/rbi/lib/openai/models/auto_file_chunking_strategy_param.rbi index 43d54634..58bd0717 100644 --- a/rbi/lib/openai/models/auto_file_chunking_strategy_param.rbi +++ b/rbi/lib/openai/models/auto_file_chunking_strategy_param.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class AutoFileChunkingStrategyParam < OpenAI::BaseModel + class AutoFileChunkingStrategyParam < OpenAI::Internal::Type::BaseModel # Always `auto`. sig { returns(Symbol) } attr_accessor :type diff --git a/rbi/lib/openai/models/batch.rbi b/rbi/lib/openai/models/batch.rbi index ddeafb8b..b34e5a45 100644 --- a/rbi/lib/openai/models/batch.rbi +++ b/rbi/lib/openai/models/batch.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class Batch < OpenAI::BaseModel + class Batch < OpenAI::Internal::Type::BaseModel sig { returns(String) } attr_accessor :id @@ -61,7 +61,7 @@ module OpenAI sig { returns(T.nilable(OpenAI::Models::Batch::Errors)) } attr_reader :errors - sig { params(errors: T.any(OpenAI::Models::Batch::Errors, OpenAI::Internal::Util::AnyHash)).void } + sig { params(errors: T.any(OpenAI::Models::Batch::Errors, OpenAI::Internal::AnyHash)).void } attr_writer :errors # The Unix timestamp (in seconds) for when the batch expired. @@ -119,7 +119,7 @@ module OpenAI sig { returns(T.nilable(OpenAI::Models::BatchRequestCounts)) } attr_reader :request_counts - sig { params(request_counts: T.any(OpenAI::Models::BatchRequestCounts, OpenAI::Internal::Util::AnyHash)).void } + sig { params(request_counts: T.any(OpenAI::Models::BatchRequestCounts, OpenAI::Internal::AnyHash)).void } attr_writer :request_counts sig do @@ -134,7 +134,7 @@ module OpenAI cancelling_at: Integer, completed_at: Integer, error_file_id: String, - errors: T.any(OpenAI::Models::Batch::Errors, OpenAI::Internal::Util::AnyHash), + errors: T.any(OpenAI::Models::Batch::Errors, OpenAI::Internal::AnyHash), expired_at: Integer, expires_at: Integer, failed_at: Integer, @@ -142,7 +142,7 @@ module OpenAI in_progress_at: Integer, metadata: T.nilable(T::Hash[Symbol, String]), output_file_id: String, - request_counts: T.any(OpenAI::Models::BatchRequestCounts, OpenAI::Internal::Util::AnyHash), + request_counts: T.any(OpenAI::Models::BatchRequestCounts, OpenAI::Internal::AnyHash), object: Symbol ) .returns(T.attached_class) @@ -203,7 +203,7 @@ module OpenAI # The current status of the batch. module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Batch::Status) } OrSymbol = T.type_alias { T.any(Symbol, String, OpenAI::Models::Batch::Status::TaggedSymbol) } @@ -222,11 +222,11 @@ module OpenAI end end - class Errors < OpenAI::BaseModel + class Errors < OpenAI::Internal::Type::BaseModel sig { returns(T.nilable(T::Array[OpenAI::Models::BatchError])) } attr_reader :data - sig { params(data: T::Array[T.any(OpenAI::Models::BatchError, OpenAI::Internal::Util::AnyHash)]).void } + sig { params(data: T::Array[T.any(OpenAI::Models::BatchError, OpenAI::Internal::AnyHash)]).void } attr_writer :data # The object type, which is always `list`. @@ -237,7 +237,7 @@ module OpenAI attr_writer :object sig do - params(data: T::Array[T.any(OpenAI::Models::BatchError, OpenAI::Internal::Util::AnyHash)], object: String) + params(data: T::Array[T.any(OpenAI::Models::BatchError, OpenAI::Internal::AnyHash)], object: String) .returns(T.attached_class) end def self.new(data: nil, object: nil) diff --git a/rbi/lib/openai/models/batch_cancel_params.rbi b/rbi/lib/openai/models/batch_cancel_params.rbi index dfdfdb51..e5ef1114 100644 --- a/rbi/lib/openai/models/batch_cancel_params.rbi +++ b/rbi/lib/openai/models/batch_cancel_params.rbi @@ -2,13 +2,17 @@ module OpenAI module Models - class BatchCancelParams < OpenAI::BaseModel + class BatchCancelParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters sig do - params(request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) - .returns(T.attached_class) + params( + request_options: T.any( + OpenAI::RequestOptions, + OpenAI::Internal::AnyHash + ) + ).returns(T.attached_class) end def self.new(request_options: {}) end diff --git a/rbi/lib/openai/models/batch_create_params.rbi b/rbi/lib/openai/models/batch_create_params.rbi index 4a38d820..d27284f7 100644 --- a/rbi/lib/openai/models/batch_create_params.rbi +++ b/rbi/lib/openai/models/batch_create_params.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class BatchCreateParams < OpenAI::BaseModel + class BatchCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -45,7 +45,7 @@ module OpenAI endpoint: OpenAI::Models::BatchCreateParams::Endpoint::OrSymbol, input_file_id: String, metadata: T.nilable(T::Hash[Symbol, String]), - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -70,7 +70,7 @@ module OpenAI # The time frame within which the batch should be processed. Currently only `24h` # is supported. module CompletionWindow - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::BatchCreateParams::CompletionWindow) } OrSymbol = @@ -88,7 +88,7 @@ module OpenAI # are supported. Note that `/v1/embeddings` batches are also restricted to a # maximum of 50,000 embedding inputs across all requests in the batch. module Endpoint - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::BatchCreateParams::Endpoint) } OrSymbol = diff --git a/rbi/lib/openai/models/batch_error.rbi b/rbi/lib/openai/models/batch_error.rbi index 261a3ac1..157787c8 100644 --- a/rbi/lib/openai/models/batch_error.rbi +++ b/rbi/lib/openai/models/batch_error.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class BatchError < OpenAI::BaseModel + class BatchError < OpenAI::Internal::Type::BaseModel # An error code identifying the error type. sig { returns(T.nilable(String)) } attr_reader :code diff --git a/rbi/lib/openai/models/batch_list_params.rbi b/rbi/lib/openai/models/batch_list_params.rbi index 498b22b2..479dcdff 100644 --- a/rbi/lib/openai/models/batch_list_params.rbi +++ b/rbi/lib/openai/models/batch_list_params.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class BatchListParams < OpenAI::BaseModel + class BatchListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -28,7 +28,7 @@ module OpenAI params( after: String, limit: Integer, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end diff --git a/rbi/lib/openai/models/batch_request_counts.rbi b/rbi/lib/openai/models/batch_request_counts.rbi index 03b6cecc..6b688106 100644 --- a/rbi/lib/openai/models/batch_request_counts.rbi +++ b/rbi/lib/openai/models/batch_request_counts.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class BatchRequestCounts < OpenAI::BaseModel + class BatchRequestCounts < OpenAI::Internal::Type::BaseModel # Number of requests that have been completed successfully. sig { returns(Integer) } attr_accessor :completed diff --git a/rbi/lib/openai/models/batch_retrieve_params.rbi b/rbi/lib/openai/models/batch_retrieve_params.rbi index ced68a22..002b5af2 100644 --- a/rbi/lib/openai/models/batch_retrieve_params.rbi +++ b/rbi/lib/openai/models/batch_retrieve_params.rbi @@ -2,13 +2,17 @@ module OpenAI module Models - class BatchRetrieveParams < OpenAI::BaseModel + class BatchRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters sig do - params(request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) - .returns(T.attached_class) + params( + request_options: T.any( + OpenAI::RequestOptions, + OpenAI::Internal::AnyHash + ) + ).returns(T.attached_class) end def self.new(request_options: {}) end diff --git a/rbi/lib/openai/models/beta/assistant.rbi b/rbi/lib/openai/models/beta/assistant.rbi index de41edc8..3ea623d1 100644 --- a/rbi/lib/openai/models/beta/assistant.rbi +++ b/rbi/lib/openai/models/beta/assistant.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Beta - class Assistant < OpenAI::BaseModel + class Assistant < OpenAI::Internal::Type::BaseModel # The identifier, which can be referenced in API endpoints. sig { returns(String) } attr_accessor :id @@ -111,7 +111,7 @@ module OpenAI sig do params( - tool_resources: T.nilable(T.any(OpenAI::Models::Beta::Assistant::ToolResources, OpenAI::Internal::Util::AnyHash)) + tool_resources: T.nilable(T.any(OpenAI::Models::Beta::Assistant::ToolResources, OpenAI::Internal::AnyHash)) ) .void end @@ -138,7 +138,7 @@ module OpenAI tools: T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::FileSearchTool, OpenAI::Models::Beta::FunctionTool ) @@ -147,13 +147,13 @@ module OpenAI T.any( Symbol, OpenAI::Models::ResponseFormatText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::ResponseFormatJSONObject, OpenAI::Models::ResponseFormatJSONSchema ) ), temperature: T.nilable(Float), - tool_resources: T.nilable(T.any(OpenAI::Models::Beta::Assistant::ToolResources, OpenAI::Internal::Util::AnyHash)), + tool_resources: T.nilable(T.any(OpenAI::Models::Beta::Assistant::ToolResources, OpenAI::Internal::AnyHash)), top_p: T.nilable(Float), object: Symbol ) @@ -212,13 +212,13 @@ module OpenAI def to_hash end - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel sig { returns(T.nilable(OpenAI::Models::Beta::Assistant::ToolResources::CodeInterpreter)) } attr_reader :code_interpreter sig do params( - code_interpreter: T.any(OpenAI::Models::Beta::Assistant::ToolResources::CodeInterpreter, OpenAI::Internal::Util::AnyHash) + code_interpreter: T.any(OpenAI::Models::Beta::Assistant::ToolResources::CodeInterpreter, OpenAI::Internal::AnyHash) ) .void end @@ -229,7 +229,7 @@ module OpenAI sig do params( - file_search: T.any(OpenAI::Models::Beta::Assistant::ToolResources::FileSearch, OpenAI::Internal::Util::AnyHash) + file_search: T.any(OpenAI::Models::Beta::Assistant::ToolResources::FileSearch, OpenAI::Internal::AnyHash) ) .void end @@ -241,8 +241,8 @@ module OpenAI # IDs. sig do params( - code_interpreter: T.any(OpenAI::Models::Beta::Assistant::ToolResources::CodeInterpreter, OpenAI::Internal::Util::AnyHash), - file_search: T.any(OpenAI::Models::Beta::Assistant::ToolResources::FileSearch, OpenAI::Internal::Util::AnyHash) + code_interpreter: T.any(OpenAI::Models::Beta::Assistant::ToolResources::CodeInterpreter, OpenAI::Internal::AnyHash), + file_search: T.any(OpenAI::Models::Beta::Assistant::ToolResources::FileSearch, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -261,7 +261,7 @@ module OpenAI def to_hash end - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel # A list of [file](https://platform.openai.com/docs/api-reference/files) IDs made # available to the `code_interpreter`` tool. There can be a maximum of 20 files # associated with the tool. @@ -280,7 +280,7 @@ module OpenAI end end - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # The ID of the # [vector store](https://platform.openai.com/docs/api-reference/vector-stores/object) # attached to this assistant. There can be a maximum of 1 vector store attached to diff --git a/rbi/lib/openai/models/beta/assistant_create_params.rbi b/rbi/lib/openai/models/beta/assistant_create_params.rbi index 19dabe18..a75932ff 100644 --- a/rbi/lib/openai/models/beta/assistant_create_params.rbi +++ b/rbi/lib/openai/models/beta/assistant_create_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Beta - class AssistantCreateParams < OpenAI::BaseModel + class AssistantCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -95,9 +95,7 @@ module OpenAI sig do params( - tool_resources: T.nilable( - T.any(OpenAI::Models::Beta::AssistantCreateParams::ToolResources, OpenAI::Internal::Util::AnyHash) - ) + tool_resources: T.nilable(T.any(OpenAI::Models::Beta::AssistantCreateParams::ToolResources, OpenAI::Internal::AnyHash)) ) .void end @@ -126,7 +124,7 @@ module OpenAI tools: T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::FileSearchTool, OpenAI::Models::Beta::FunctionTool ) @@ -156,25 +154,23 @@ module OpenAI T.any( Symbol, OpenAI::Models::ResponseFormatText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::ResponseFormatJSONObject, OpenAI::Models::ResponseFormatJSONSchema ) ), temperature: T.nilable(Float), - tool_resources: T.nilable( - T.any(OpenAI::Models::Beta::AssistantCreateParams::ToolResources, OpenAI::Internal::Util::AnyHash) - ), + tool_resources: T.nilable(T.any(OpenAI::Models::Beta::AssistantCreateParams::ToolResources, OpenAI::Internal::AnyHash)), tools: T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::FileSearchTool, OpenAI::Models::Beta::FunctionTool ) ], top_p: T.nilable(Float), - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -235,14 +231,14 @@ module OpenAI # [Model overview](https://platform.openai.com/docs/models) for descriptions of # them. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, OpenAI::Models::ChatModel::OrSymbol]) } def self.variants end end - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel sig { returns(T.nilable(OpenAI::Models::Beta::AssistantCreateParams::ToolResources::CodeInterpreter)) } attr_reader :code_interpreter @@ -250,7 +246,7 @@ module OpenAI params( code_interpreter: T.any( OpenAI::Models::Beta::AssistantCreateParams::ToolResources::CodeInterpreter, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .void @@ -262,10 +258,7 @@ module OpenAI sig do params( - file_search: T.any( - OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch, - OpenAI::Internal::Util::AnyHash - ) + file_search: T.any(OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch, OpenAI::Internal::AnyHash) ) .void end @@ -279,12 +272,9 @@ module OpenAI params( code_interpreter: T.any( OpenAI::Models::Beta::AssistantCreateParams::ToolResources::CodeInterpreter, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ), - file_search: T.any( - OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch, - OpenAI::Internal::Util::AnyHash - ) + file_search: T.any(OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -303,7 +293,7 @@ module OpenAI def to_hash end - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel # A list of [file](https://platform.openai.com/docs/api-reference/files) IDs made # available to the `code_interpreter` tool. There can be a maximum of 20 files # associated with the tool. @@ -322,7 +312,7 @@ module OpenAI end end - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # The # [vector store](https://platform.openai.com/docs/api-reference/vector-stores/object) # attached to this assistant. There can be a maximum of 1 vector store attached to @@ -349,7 +339,7 @@ module OpenAI vector_stores: T::Array[ T.any( OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch::VectorStore, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ] ) @@ -363,7 +353,7 @@ module OpenAI vector_stores: T::Array[ T.any( OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch::VectorStore, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ] ) @@ -384,7 +374,7 @@ module OpenAI def to_hash end - class VectorStore < OpenAI::BaseModel + class VectorStore < OpenAI::Internal::Type::BaseModel # The chunking strategy used to chunk the file(s). If not set, will use the `auto` # strategy. sig do @@ -403,7 +393,7 @@ module OpenAI params( chunking_strategy: T.any( OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Auto, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static ) ) @@ -433,7 +423,7 @@ module OpenAI params( chunking_strategy: T.any( OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Auto, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static ), file_ids: T::Array[String], @@ -463,9 +453,9 @@ module OpenAI # The chunking strategy used to chunk the file(s). If not set, will use the `auto` # strategy. module ChunkingStrategy - extend OpenAI::Union + extend OpenAI::Internal::Type::Union - class Auto < OpenAI::BaseModel + class Auto < OpenAI::Internal::Type::BaseModel # Always `auto`. sig { returns(Symbol) } attr_accessor :type @@ -481,7 +471,7 @@ module OpenAI end end - class Static < OpenAI::BaseModel + class Static < OpenAI::Internal::Type::BaseModel sig do returns( OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static::Static @@ -493,7 +483,7 @@ module OpenAI params( static: T.any( OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static::Static, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .void @@ -508,7 +498,7 @@ module OpenAI params( static: T.any( OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static::Static, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ), type: Symbol ) @@ -529,7 +519,7 @@ module OpenAI def to_hash end - class Static < OpenAI::BaseModel + class Static < OpenAI::Internal::Type::BaseModel # The number of tokens that overlap between chunks. The default value is `400`. # # Note that the overlap must not exceed half of `max_chunk_size_tokens`. diff --git a/rbi/lib/openai/models/beta/assistant_delete_params.rbi b/rbi/lib/openai/models/beta/assistant_delete_params.rbi index 31fbc46f..b929b1f8 100644 --- a/rbi/lib/openai/models/beta/assistant_delete_params.rbi +++ b/rbi/lib/openai/models/beta/assistant_delete_params.rbi @@ -3,13 +3,17 @@ module OpenAI module Models module Beta - class AssistantDeleteParams < OpenAI::BaseModel + class AssistantDeleteParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters sig do - params(request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) - .returns(T.attached_class) + params( + request_options: T.any( + OpenAI::RequestOptions, + OpenAI::Internal::AnyHash + ) + ).returns(T.attached_class) end def self.new(request_options: {}) end diff --git a/rbi/lib/openai/models/beta/assistant_deleted.rbi b/rbi/lib/openai/models/beta/assistant_deleted.rbi index c5b56819..08a9c254 100644 --- a/rbi/lib/openai/models/beta/assistant_deleted.rbi +++ b/rbi/lib/openai/models/beta/assistant_deleted.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Beta - class AssistantDeleted < OpenAI::BaseModel + class AssistantDeleted < OpenAI::Internal::Type::BaseModel sig { returns(String) } attr_accessor :id diff --git a/rbi/lib/openai/models/beta/assistant_list_params.rbi b/rbi/lib/openai/models/beta/assistant_list_params.rbi index c8a24c36..7da7715a 100644 --- a/rbi/lib/openai/models/beta/assistant_list_params.rbi +++ b/rbi/lib/openai/models/beta/assistant_list_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Beta - class AssistantListParams < OpenAI::BaseModel + class AssistantListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -49,7 +49,7 @@ module OpenAI before: String, limit: Integer, order: OpenAI::Models::Beta::AssistantListParams::Order::OrSymbol, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -74,7 +74,7 @@ module OpenAI # Sort order by the `created_at` timestamp of the objects. `asc` for ascending # order and `desc` for descending order. module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::AssistantListParams::Order) } OrSymbol = diff --git a/rbi/lib/openai/models/beta/assistant_response_format_option.rbi b/rbi/lib/openai/models/beta/assistant_response_format_option.rbi index a44a0adc..0b95306f 100644 --- a/rbi/lib/openai/models/beta/assistant_response_format_option.rbi +++ b/rbi/lib/openai/models/beta/assistant_response_format_option.rbi @@ -24,7 +24,7 @@ module OpenAI # indicates the generation exceeded `max_tokens` or the conversation exceeded the # max context length. module AssistantResponseFormatOption - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/beta/assistant_retrieve_params.rbi b/rbi/lib/openai/models/beta/assistant_retrieve_params.rbi index 3e5121e0..f118b741 100644 --- a/rbi/lib/openai/models/beta/assistant_retrieve_params.rbi +++ b/rbi/lib/openai/models/beta/assistant_retrieve_params.rbi @@ -3,13 +3,17 @@ module OpenAI module Models module Beta - class AssistantRetrieveParams < OpenAI::BaseModel + class AssistantRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters sig do - params(request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) - .returns(T.attached_class) + params( + request_options: T.any( + OpenAI::RequestOptions, + OpenAI::Internal::AnyHash + ) + ).returns(T.attached_class) end def self.new(request_options: {}) end diff --git a/rbi/lib/openai/models/beta/assistant_stream_event.rbi b/rbi/lib/openai/models/beta/assistant_stream_event.rbi index e3b90b3b..86e113fd 100644 --- a/rbi/lib/openai/models/beta/assistant_stream_event.rbi +++ b/rbi/lib/openai/models/beta/assistant_stream_event.rbi @@ -24,15 +24,15 @@ module OpenAI # [Assistants API quickstart](https://platform.openai.com/docs/assistants/overview) # to learn how to integrate the Assistants API with streaming. module AssistantStreamEvent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union - class ThreadCreated < OpenAI::BaseModel + class ThreadCreated < OpenAI::Internal::Type::BaseModel # Represents a thread that contains # [messages](https://platform.openai.com/docs/api-reference/messages). sig { returns(OpenAI::Models::Beta::Thread) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Thread, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Thread, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -50,7 +50,7 @@ module OpenAI # created. sig do params( - data: T.any(OpenAI::Models::Beta::Thread, OpenAI::Internal::Util::AnyHash), + data: T.any(OpenAI::Models::Beta::Thread, OpenAI::Internal::AnyHash), enabled: T::Boolean, event: Symbol ) @@ -64,13 +64,13 @@ module OpenAI end end - class ThreadRunCreated < OpenAI::BaseModel + class ThreadRunCreated < OpenAI::Internal::Type::BaseModel # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Run) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -79,7 +79,7 @@ module OpenAI # Occurs when a new # [run](https://platform.openai.com/docs/api-reference/runs/object) is created. sig do - params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash), event: Symbol) + params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.run.created") @@ -90,13 +90,13 @@ module OpenAI end end - class ThreadRunQueued < OpenAI::BaseModel + class ThreadRunQueued < OpenAI::Internal::Type::BaseModel # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Run) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -105,7 +105,7 @@ module OpenAI # Occurs when a [run](https://platform.openai.com/docs/api-reference/runs/object) # moves to a `queued` status. sig do - params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash), event: Symbol) + params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.run.queued") @@ -116,13 +116,13 @@ module OpenAI end end - class ThreadRunInProgress < OpenAI::BaseModel + class ThreadRunInProgress < OpenAI::Internal::Type::BaseModel # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Run) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -131,7 +131,7 @@ module OpenAI # Occurs when a [run](https://platform.openai.com/docs/api-reference/runs/object) # moves to an `in_progress` status. sig do - params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash), event: Symbol) + params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.run.in_progress") @@ -142,13 +142,13 @@ module OpenAI end end - class ThreadRunRequiresAction < OpenAI::BaseModel + class ThreadRunRequiresAction < OpenAI::Internal::Type::BaseModel # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Run) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -157,7 +157,7 @@ module OpenAI # Occurs when a [run](https://platform.openai.com/docs/api-reference/runs/object) # moves to a `requires_action` status. sig do - params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash), event: Symbol) + params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.run.requires_action") @@ -168,13 +168,13 @@ module OpenAI end end - class ThreadRunCompleted < OpenAI::BaseModel + class ThreadRunCompleted < OpenAI::Internal::Type::BaseModel # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Run) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -183,7 +183,7 @@ module OpenAI # Occurs when a [run](https://platform.openai.com/docs/api-reference/runs/object) # is completed. sig do - params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash), event: Symbol) + params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.run.completed") @@ -194,13 +194,13 @@ module OpenAI end end - class ThreadRunIncomplete < OpenAI::BaseModel + class ThreadRunIncomplete < OpenAI::Internal::Type::BaseModel # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Run) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -209,7 +209,7 @@ module OpenAI # Occurs when a [run](https://platform.openai.com/docs/api-reference/runs/object) # ends with status `incomplete`. sig do - params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash), event: Symbol) + params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.run.incomplete") @@ -220,13 +220,13 @@ module OpenAI end end - class ThreadRunFailed < OpenAI::BaseModel + class ThreadRunFailed < OpenAI::Internal::Type::BaseModel # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Run) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -235,7 +235,7 @@ module OpenAI # Occurs when a [run](https://platform.openai.com/docs/api-reference/runs/object) # fails. sig do - params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash), event: Symbol) + params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.run.failed") @@ -246,13 +246,13 @@ module OpenAI end end - class ThreadRunCancelling < OpenAI::BaseModel + class ThreadRunCancelling < OpenAI::Internal::Type::BaseModel # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Run) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -261,7 +261,7 @@ module OpenAI # Occurs when a [run](https://platform.openai.com/docs/api-reference/runs/object) # moves to a `cancelling` status. sig do - params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash), event: Symbol) + params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.run.cancelling") @@ -272,13 +272,13 @@ module OpenAI end end - class ThreadRunCancelled < OpenAI::BaseModel + class ThreadRunCancelled < OpenAI::Internal::Type::BaseModel # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Run) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -287,7 +287,7 @@ module OpenAI # Occurs when a [run](https://platform.openai.com/docs/api-reference/runs/object) # is cancelled. sig do - params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash), event: Symbol) + params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.run.cancelled") @@ -298,13 +298,13 @@ module OpenAI end end - class ThreadRunExpired < OpenAI::BaseModel + class ThreadRunExpired < OpenAI::Internal::Type::BaseModel # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Run) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -313,7 +313,7 @@ module OpenAI # Occurs when a [run](https://platform.openai.com/docs/api-reference/runs/object) # expires. sig do - params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash), event: Symbol) + params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.run.expired") @@ -324,12 +324,12 @@ module OpenAI end end - class ThreadRunStepCreated < OpenAI::BaseModel + class ThreadRunStepCreated < OpenAI::Internal::Type::BaseModel # Represents a step in execution of a run. sig { returns(OpenAI::Models::Beta::Threads::Runs::RunStep) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -340,7 +340,7 @@ module OpenAI # is created. sig do params( - data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash), + data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash), event: Symbol ) .returns(T.attached_class) @@ -353,12 +353,12 @@ module OpenAI end end - class ThreadRunStepInProgress < OpenAI::BaseModel + class ThreadRunStepInProgress < OpenAI::Internal::Type::BaseModel # Represents a step in execution of a run. sig { returns(OpenAI::Models::Beta::Threads::Runs::RunStep) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -369,7 +369,7 @@ module OpenAI # moves to an `in_progress` state. sig do params( - data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash), + data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash), event: Symbol ) .returns(T.attached_class) @@ -382,18 +382,13 @@ module OpenAI end end - class ThreadRunStepDelta < OpenAI::BaseModel + class ThreadRunStepDelta < OpenAI::Internal::Type::BaseModel # Represents a run step delta i.e. any changed fields on a run step during # streaming. sig { returns(OpenAI::Models::Beta::Threads::Runs::RunStepDeltaEvent) } attr_reader :data - sig do - params( - data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStepDeltaEvent, OpenAI::Internal::Util::AnyHash) - ) - .void - end + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStepDeltaEvent, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -404,7 +399,7 @@ module OpenAI # are being streamed. sig do params( - data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStepDeltaEvent, OpenAI::Internal::Util::AnyHash), + data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStepDeltaEvent, OpenAI::Internal::AnyHash), event: Symbol ) .returns(T.attached_class) @@ -417,12 +412,12 @@ module OpenAI end end - class ThreadRunStepCompleted < OpenAI::BaseModel + class ThreadRunStepCompleted < OpenAI::Internal::Type::BaseModel # Represents a step in execution of a run. sig { returns(OpenAI::Models::Beta::Threads::Runs::RunStep) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -433,7 +428,7 @@ module OpenAI # is completed. sig do params( - data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash), + data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash), event: Symbol ) .returns(T.attached_class) @@ -446,12 +441,12 @@ module OpenAI end end - class ThreadRunStepFailed < OpenAI::BaseModel + class ThreadRunStepFailed < OpenAI::Internal::Type::BaseModel # Represents a step in execution of a run. sig { returns(OpenAI::Models::Beta::Threads::Runs::RunStep) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -462,7 +457,7 @@ module OpenAI # fails. sig do params( - data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash), + data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash), event: Symbol ) .returns(T.attached_class) @@ -475,12 +470,12 @@ module OpenAI end end - class ThreadRunStepCancelled < OpenAI::BaseModel + class ThreadRunStepCancelled < OpenAI::Internal::Type::BaseModel # Represents a step in execution of a run. sig { returns(OpenAI::Models::Beta::Threads::Runs::RunStep) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -491,7 +486,7 @@ module OpenAI # is cancelled. sig do params( - data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash), + data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash), event: Symbol ) .returns(T.attached_class) @@ -504,12 +499,12 @@ module OpenAI end end - class ThreadRunStepExpired < OpenAI::BaseModel + class ThreadRunStepExpired < OpenAI::Internal::Type::BaseModel # Represents a step in execution of a run. sig { returns(OpenAI::Models::Beta::Threads::Runs::RunStep) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -520,7 +515,7 @@ module OpenAI # expires. sig do params( - data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash), + data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash), event: Symbol ) .returns(T.attached_class) @@ -533,13 +528,13 @@ module OpenAI end end - class ThreadMessageCreated < OpenAI::BaseModel + class ThreadMessageCreated < OpenAI::Internal::Type::BaseModel # Represents a message within a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Message) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -549,10 +544,7 @@ module OpenAI # [message](https://platform.openai.com/docs/api-reference/messages/object) is # created. sig do - params( - data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::Util::AnyHash), - event: Symbol - ) + params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.message.created") @@ -563,13 +555,13 @@ module OpenAI end end - class ThreadMessageInProgress < OpenAI::BaseModel + class ThreadMessageInProgress < OpenAI::Internal::Type::BaseModel # Represents a message within a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Message) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -579,10 +571,7 @@ module OpenAI # [message](https://platform.openai.com/docs/api-reference/messages/object) moves # to an `in_progress` state. sig do - params( - data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::Util::AnyHash), - event: Symbol - ) + params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.message.in_progress") @@ -593,13 +582,13 @@ module OpenAI end end - class ThreadMessageDelta < OpenAI::BaseModel + class ThreadMessageDelta < OpenAI::Internal::Type::BaseModel # Represents a message delta i.e. any changed fields on a message during # streaming. sig { returns(OpenAI::Models::Beta::Threads::MessageDeltaEvent) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::MessageDeltaEvent, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::MessageDeltaEvent, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -610,7 +599,7 @@ module OpenAI # being streamed. sig do params( - data: T.any(OpenAI::Models::Beta::Threads::MessageDeltaEvent, OpenAI::Internal::Util::AnyHash), + data: T.any(OpenAI::Models::Beta::Threads::MessageDeltaEvent, OpenAI::Internal::AnyHash), event: Symbol ) .returns(T.attached_class) @@ -623,13 +612,13 @@ module OpenAI end end - class ThreadMessageCompleted < OpenAI::BaseModel + class ThreadMessageCompleted < OpenAI::Internal::Type::BaseModel # Represents a message within a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Message) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -639,10 +628,7 @@ module OpenAI # [message](https://platform.openai.com/docs/api-reference/messages/object) is # completed. sig do - params( - data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::Util::AnyHash), - event: Symbol - ) + params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.message.completed") @@ -653,13 +639,13 @@ module OpenAI end end - class ThreadMessageIncomplete < OpenAI::BaseModel + class ThreadMessageIncomplete < OpenAI::Internal::Type::BaseModel # Represents a message within a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Message) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -669,10 +655,7 @@ module OpenAI # [message](https://platform.openai.com/docs/api-reference/messages/object) ends # before it is completed. sig do - params( - data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::Util::AnyHash), - event: Symbol - ) + params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.message.incomplete") @@ -683,11 +666,11 @@ module OpenAI end end - class ErrorEvent < OpenAI::BaseModel + class ErrorEvent < OpenAI::Internal::Type::BaseModel sig { returns(OpenAI::Models::ErrorObject) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::ErrorObject, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::ErrorObject, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -697,7 +680,7 @@ module OpenAI # [error](https://platform.openai.com/docs/guides/error-codes#api-errors) occurs. # This can happen due to an internal server error or a timeout. sig do - params(data: T.any(OpenAI::Models::ErrorObject, OpenAI::Internal::Util::AnyHash), event: Symbol) + params(data: T.any(OpenAI::Models::ErrorObject, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :error) diff --git a/rbi/lib/openai/models/beta/assistant_tool.rbi b/rbi/lib/openai/models/beta/assistant_tool.rbi index f9ca4d29..e2eaccc9 100644 --- a/rbi/lib/openai/models/beta/assistant_tool.rbi +++ b/rbi/lib/openai/models/beta/assistant_tool.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module AssistantTool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/beta/assistant_tool_choice.rbi b/rbi/lib/openai/models/beta/assistant_tool_choice.rbi index 36c345d4..79dcf2c8 100644 --- a/rbi/lib/openai/models/beta/assistant_tool_choice.rbi +++ b/rbi/lib/openai/models/beta/assistant_tool_choice.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Beta - class AssistantToolChoice < OpenAI::BaseModel + class AssistantToolChoice < OpenAI::Internal::Type::BaseModel # The type of the tool. If type is `function`, the function name must be set sig { returns(OpenAI::Models::Beta::AssistantToolChoice::Type::OrSymbol) } attr_accessor :type @@ -11,12 +11,7 @@ module OpenAI sig { returns(T.nilable(OpenAI::Models::Beta::AssistantToolChoiceFunction)) } attr_reader :function - sig do - params( - function: T.any(OpenAI::Models::Beta::AssistantToolChoiceFunction, OpenAI::Internal::Util::AnyHash) - ) - .void - end + sig { params(function: T.any(OpenAI::Models::Beta::AssistantToolChoiceFunction, OpenAI::Internal::AnyHash)).void } attr_writer :function # Specifies a tool the model should use. Use to force the model to call a specific @@ -24,7 +19,7 @@ module OpenAI sig do params( type: OpenAI::Models::Beta::AssistantToolChoice::Type::OrSymbol, - function: T.any(OpenAI::Models::Beta::AssistantToolChoiceFunction, OpenAI::Internal::Util::AnyHash) + function: T.any(OpenAI::Models::Beta::AssistantToolChoiceFunction, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -45,7 +40,7 @@ module OpenAI # The type of the tool. If type is `function`, the function name must be set module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::AssistantToolChoice::Type) } OrSymbol = diff --git a/rbi/lib/openai/models/beta/assistant_tool_choice_function.rbi b/rbi/lib/openai/models/beta/assistant_tool_choice_function.rbi index 084e57a8..bbf08c85 100644 --- a/rbi/lib/openai/models/beta/assistant_tool_choice_function.rbi +++ b/rbi/lib/openai/models/beta/assistant_tool_choice_function.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Beta - class AssistantToolChoiceFunction < OpenAI::BaseModel + class AssistantToolChoiceFunction < OpenAI::Internal::Type::BaseModel # The name of the function to call. sig { returns(String) } attr_accessor :name diff --git a/rbi/lib/openai/models/beta/assistant_tool_choice_option.rbi b/rbi/lib/openai/models/beta/assistant_tool_choice_option.rbi index 44079027..3a3a1f7c 100644 --- a/rbi/lib/openai/models/beta/assistant_tool_choice_option.rbi +++ b/rbi/lib/openai/models/beta/assistant_tool_choice_option.rbi @@ -11,14 +11,14 @@ module OpenAI # `{"type": "function", "function": {"name": "my_function"}}` forces the model to # call that tool. module AssistantToolChoiceOption - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # `none` means the model will not call any tools and instead generates a message. # `auto` means the model can pick between generating a message or calling one or # more tools. `required` means the model must call one or more tools before # responding to the user. module Auto - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::AssistantToolChoiceOption::Auto) } OrSymbol = diff --git a/rbi/lib/openai/models/beta/assistant_update_params.rbi b/rbi/lib/openai/models/beta/assistant_update_params.rbi index 929c5b35..f14621d3 100644 --- a/rbi/lib/openai/models/beta/assistant_update_params.rbi +++ b/rbi/lib/openai/models/beta/assistant_update_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Beta - class AssistantUpdateParams < OpenAI::BaseModel + class AssistantUpdateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -98,9 +98,7 @@ module OpenAI sig do params( - tool_resources: T.nilable( - T.any(OpenAI::Models::Beta::AssistantUpdateParams::ToolResources, OpenAI::Internal::Util::AnyHash) - ) + tool_resources: T.nilable(T.any(OpenAI::Models::Beta::AssistantUpdateParams::ToolResources, OpenAI::Internal::AnyHash)) ) .void end @@ -129,7 +127,7 @@ module OpenAI tools: T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::FileSearchTool, OpenAI::Models::Beta::FunctionTool ) @@ -159,25 +157,23 @@ module OpenAI T.any( Symbol, OpenAI::Models::ResponseFormatText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::ResponseFormatJSONObject, OpenAI::Models::ResponseFormatJSONSchema ) ), temperature: T.nilable(Float), - tool_resources: T.nilable( - T.any(OpenAI::Models::Beta::AssistantUpdateParams::ToolResources, OpenAI::Internal::Util::AnyHash) - ), + tool_resources: T.nilable(T.any(OpenAI::Models::Beta::AssistantUpdateParams::ToolResources, OpenAI::Internal::AnyHash)), tools: T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::FileSearchTool, OpenAI::Models::Beta::FunctionTool ) ], top_p: T.nilable(Float), - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -238,7 +234,7 @@ module OpenAI # [Model overview](https://platform.openai.com/docs/models) for descriptions of # them. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, OpenAI::Models::Beta::AssistantUpdateParams::Model::OrSymbol]) } def self.variants @@ -299,7 +295,7 @@ module OpenAI T.let(:"gpt-3.5-turbo-16k-0613", OpenAI::Models::Beta::AssistantUpdateParams::Model::TaggedSymbol) end - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel sig { returns(T.nilable(OpenAI::Models::Beta::AssistantUpdateParams::ToolResources::CodeInterpreter)) } attr_reader :code_interpreter @@ -307,7 +303,7 @@ module OpenAI params( code_interpreter: T.any( OpenAI::Models::Beta::AssistantUpdateParams::ToolResources::CodeInterpreter, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .void @@ -319,10 +315,7 @@ module OpenAI sig do params( - file_search: T.any( - OpenAI::Models::Beta::AssistantUpdateParams::ToolResources::FileSearch, - OpenAI::Internal::Util::AnyHash - ) + file_search: T.any(OpenAI::Models::Beta::AssistantUpdateParams::ToolResources::FileSearch, OpenAI::Internal::AnyHash) ) .void end @@ -336,12 +329,9 @@ module OpenAI params( code_interpreter: T.any( OpenAI::Models::Beta::AssistantUpdateParams::ToolResources::CodeInterpreter, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ), - file_search: T.any( - OpenAI::Models::Beta::AssistantUpdateParams::ToolResources::FileSearch, - OpenAI::Internal::Util::AnyHash - ) + file_search: T.any(OpenAI::Models::Beta::AssistantUpdateParams::ToolResources::FileSearch, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -360,7 +350,7 @@ module OpenAI def to_hash end - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel # Overrides the list of # [file](https://platform.openai.com/docs/api-reference/files) IDs made available # to the `code_interpreter` tool. There can be a maximum of 20 files associated @@ -380,7 +370,7 @@ module OpenAI end end - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # Overrides the # [vector store](https://platform.openai.com/docs/api-reference/vector-stores/object) # attached to this assistant. There can be a maximum of 1 vector store attached to diff --git a/rbi/lib/openai/models/beta/code_interpreter_tool.rbi b/rbi/lib/openai/models/beta/code_interpreter_tool.rbi index 3e68afee..f629a549 100644 --- a/rbi/lib/openai/models/beta/code_interpreter_tool.rbi +++ b/rbi/lib/openai/models/beta/code_interpreter_tool.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Beta - class CodeInterpreterTool < OpenAI::BaseModel + class CodeInterpreterTool < OpenAI::Internal::Type::BaseModel # The type of tool being defined: `code_interpreter` sig { returns(Symbol) } attr_accessor :type diff --git a/rbi/lib/openai/models/beta/file_search_tool.rbi b/rbi/lib/openai/models/beta/file_search_tool.rbi index e13ceaf5..7388dc69 100644 --- a/rbi/lib/openai/models/beta/file_search_tool.rbi +++ b/rbi/lib/openai/models/beta/file_search_tool.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Beta - class FileSearchTool < OpenAI::BaseModel + class FileSearchTool < OpenAI::Internal::Type::BaseModel # The type of tool being defined: `file_search` sig { returns(Symbol) } attr_accessor :type @@ -13,16 +13,14 @@ module OpenAI attr_reader :file_search sig do - params( - file_search: T.any(OpenAI::Models::Beta::FileSearchTool::FileSearch, OpenAI::Internal::Util::AnyHash) - ) + params(file_search: T.any(OpenAI::Models::Beta::FileSearchTool::FileSearch, OpenAI::Internal::AnyHash)) .void end attr_writer :file_search sig do params( - file_search: T.any(OpenAI::Models::Beta::FileSearchTool::FileSearch, OpenAI::Internal::Util::AnyHash), + file_search: T.any(OpenAI::Models::Beta::FileSearchTool::FileSearch, OpenAI::Internal::AnyHash), type: Symbol ) .returns(T.attached_class) @@ -34,7 +32,7 @@ module OpenAI def to_hash end - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # The maximum number of results the file search tool should output. The default is # 20 for `gpt-4*` models and 5 for `gpt-3.5-turbo`. This number should be between # 1 and 50 inclusive. @@ -60,7 +58,7 @@ module OpenAI sig do params( - ranking_options: T.any(OpenAI::Models::Beta::FileSearchTool::FileSearch::RankingOptions, OpenAI::Internal::Util::AnyHash) + ranking_options: T.any(OpenAI::Models::Beta::FileSearchTool::FileSearch::RankingOptions, OpenAI::Internal::AnyHash) ) .void end @@ -70,7 +68,7 @@ module OpenAI sig do params( max_num_results: Integer, - ranking_options: T.any(OpenAI::Models::Beta::FileSearchTool::FileSearch::RankingOptions, OpenAI::Internal::Util::AnyHash) + ranking_options: T.any(OpenAI::Models::Beta::FileSearchTool::FileSearch::RankingOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -86,7 +84,7 @@ module OpenAI def to_hash end - class RankingOptions < OpenAI::BaseModel + class RankingOptions < OpenAI::Internal::Type::BaseModel # The score threshold for the file search. All values must be a floating point # number between 0 and 1. sig { returns(Float) } @@ -131,7 +129,7 @@ module OpenAI # The ranker to use for the file search. If not specified will use the `auto` # ranker. module Ranker - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::FileSearchTool::FileSearch::RankingOptions::Ranker) } diff --git a/rbi/lib/openai/models/beta/function_tool.rbi b/rbi/lib/openai/models/beta/function_tool.rbi index 8d337d24..c1b5305c 100644 --- a/rbi/lib/openai/models/beta/function_tool.rbi +++ b/rbi/lib/openai/models/beta/function_tool.rbi @@ -3,11 +3,11 @@ module OpenAI module Models module Beta - class FunctionTool < OpenAI::BaseModel + class FunctionTool < OpenAI::Internal::Type::BaseModel sig { returns(OpenAI::Models::FunctionDefinition) } attr_reader :function - sig { params(function: T.any(OpenAI::Models::FunctionDefinition, OpenAI::Internal::Util::AnyHash)).void } + sig { params(function: T.any(OpenAI::Models::FunctionDefinition, OpenAI::Internal::AnyHash)).void } attr_writer :function # The type of tool being defined: `function` @@ -15,7 +15,7 @@ module OpenAI attr_accessor :type sig do - params(function: T.any(OpenAI::Models::FunctionDefinition, OpenAI::Internal::Util::AnyHash), type: Symbol) + params(function: T.any(OpenAI::Models::FunctionDefinition, OpenAI::Internal::AnyHash), type: Symbol) .returns(T.attached_class) end def self.new(function:, type: :function) diff --git a/rbi/lib/openai/models/beta/message_stream_event.rbi b/rbi/lib/openai/models/beta/message_stream_event.rbi index 0c56b760..fe25c2b0 100644 --- a/rbi/lib/openai/models/beta/message_stream_event.rbi +++ b/rbi/lib/openai/models/beta/message_stream_event.rbi @@ -7,15 +7,15 @@ module OpenAI # [message](https://platform.openai.com/docs/api-reference/messages/object) is # created. module MessageStreamEvent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union - class ThreadMessageCreated < OpenAI::BaseModel + class ThreadMessageCreated < OpenAI::Internal::Type::BaseModel # Represents a message within a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Message) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -25,10 +25,7 @@ module OpenAI # [message](https://platform.openai.com/docs/api-reference/messages/object) is # created. sig do - params( - data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::Util::AnyHash), - event: Symbol - ) + params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.message.created") @@ -39,13 +36,13 @@ module OpenAI end end - class ThreadMessageInProgress < OpenAI::BaseModel + class ThreadMessageInProgress < OpenAI::Internal::Type::BaseModel # Represents a message within a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Message) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -55,10 +52,7 @@ module OpenAI # [message](https://platform.openai.com/docs/api-reference/messages/object) moves # to an `in_progress` state. sig do - params( - data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::Util::AnyHash), - event: Symbol - ) + params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.message.in_progress") @@ -69,13 +63,13 @@ module OpenAI end end - class ThreadMessageDelta < OpenAI::BaseModel + class ThreadMessageDelta < OpenAI::Internal::Type::BaseModel # Represents a message delta i.e. any changed fields on a message during # streaming. sig { returns(OpenAI::Models::Beta::Threads::MessageDeltaEvent) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::MessageDeltaEvent, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::MessageDeltaEvent, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -86,7 +80,7 @@ module OpenAI # being streamed. sig do params( - data: T.any(OpenAI::Models::Beta::Threads::MessageDeltaEvent, OpenAI::Internal::Util::AnyHash), + data: T.any(OpenAI::Models::Beta::Threads::MessageDeltaEvent, OpenAI::Internal::AnyHash), event: Symbol ) .returns(T.attached_class) @@ -99,13 +93,13 @@ module OpenAI end end - class ThreadMessageCompleted < OpenAI::BaseModel + class ThreadMessageCompleted < OpenAI::Internal::Type::BaseModel # Represents a message within a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Message) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -115,10 +109,7 @@ module OpenAI # [message](https://platform.openai.com/docs/api-reference/messages/object) is # completed. sig do - params( - data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::Util::AnyHash), - event: Symbol - ) + params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.message.completed") @@ -129,13 +120,13 @@ module OpenAI end end - class ThreadMessageIncomplete < OpenAI::BaseModel + class ThreadMessageIncomplete < OpenAI::Internal::Type::BaseModel # Represents a message within a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Message) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -145,10 +136,7 @@ module OpenAI # [message](https://platform.openai.com/docs/api-reference/messages/object) ends # before it is completed. sig do - params( - data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::Util::AnyHash), - event: Symbol - ) + params(data: T.any(OpenAI::Models::Beta::Threads::Message, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.message.incomplete") diff --git a/rbi/lib/openai/models/beta/run_step_stream_event.rbi b/rbi/lib/openai/models/beta/run_step_stream_event.rbi index dddd0b7e..0b610488 100644 --- a/rbi/lib/openai/models/beta/run_step_stream_event.rbi +++ b/rbi/lib/openai/models/beta/run_step_stream_event.rbi @@ -7,14 +7,14 @@ module OpenAI # [run step](https://platform.openai.com/docs/api-reference/run-steps/step-object) # is created. module RunStepStreamEvent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union - class ThreadRunStepCreated < OpenAI::BaseModel + class ThreadRunStepCreated < OpenAI::Internal::Type::BaseModel # Represents a step in execution of a run. sig { returns(OpenAI::Models::Beta::Threads::Runs::RunStep) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -25,7 +25,7 @@ module OpenAI # is created. sig do params( - data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash), + data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash), event: Symbol ) .returns(T.attached_class) @@ -38,12 +38,12 @@ module OpenAI end end - class ThreadRunStepInProgress < OpenAI::BaseModel + class ThreadRunStepInProgress < OpenAI::Internal::Type::BaseModel # Represents a step in execution of a run. sig { returns(OpenAI::Models::Beta::Threads::Runs::RunStep) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -54,7 +54,7 @@ module OpenAI # moves to an `in_progress` state. sig do params( - data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash), + data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash), event: Symbol ) .returns(T.attached_class) @@ -67,18 +67,13 @@ module OpenAI end end - class ThreadRunStepDelta < OpenAI::BaseModel + class ThreadRunStepDelta < OpenAI::Internal::Type::BaseModel # Represents a run step delta i.e. any changed fields on a run step during # streaming. sig { returns(OpenAI::Models::Beta::Threads::Runs::RunStepDeltaEvent) } attr_reader :data - sig do - params( - data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStepDeltaEvent, OpenAI::Internal::Util::AnyHash) - ) - .void - end + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStepDeltaEvent, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -89,7 +84,7 @@ module OpenAI # are being streamed. sig do params( - data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStepDeltaEvent, OpenAI::Internal::Util::AnyHash), + data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStepDeltaEvent, OpenAI::Internal::AnyHash), event: Symbol ) .returns(T.attached_class) @@ -102,12 +97,12 @@ module OpenAI end end - class ThreadRunStepCompleted < OpenAI::BaseModel + class ThreadRunStepCompleted < OpenAI::Internal::Type::BaseModel # Represents a step in execution of a run. sig { returns(OpenAI::Models::Beta::Threads::Runs::RunStep) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -118,7 +113,7 @@ module OpenAI # is completed. sig do params( - data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash), + data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash), event: Symbol ) .returns(T.attached_class) @@ -131,12 +126,12 @@ module OpenAI end end - class ThreadRunStepFailed < OpenAI::BaseModel + class ThreadRunStepFailed < OpenAI::Internal::Type::BaseModel # Represents a step in execution of a run. sig { returns(OpenAI::Models::Beta::Threads::Runs::RunStep) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -147,7 +142,7 @@ module OpenAI # fails. sig do params( - data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash), + data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash), event: Symbol ) .returns(T.attached_class) @@ -160,12 +155,12 @@ module OpenAI end end - class ThreadRunStepCancelled < OpenAI::BaseModel + class ThreadRunStepCancelled < OpenAI::Internal::Type::BaseModel # Represents a step in execution of a run. sig { returns(OpenAI::Models::Beta::Threads::Runs::RunStep) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -176,7 +171,7 @@ module OpenAI # is cancelled. sig do params( - data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash), + data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash), event: Symbol ) .returns(T.attached_class) @@ -189,12 +184,12 @@ module OpenAI end end - class ThreadRunStepExpired < OpenAI::BaseModel + class ThreadRunStepExpired < OpenAI::Internal::Type::BaseModel # Represents a step in execution of a run. sig { returns(OpenAI::Models::Beta::Threads::Runs::RunStep) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -205,7 +200,7 @@ module OpenAI # expires. sig do params( - data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::Util::AnyHash), + data: T.any(OpenAI::Models::Beta::Threads::Runs::RunStep, OpenAI::Internal::AnyHash), event: Symbol ) .returns(T.attached_class) diff --git a/rbi/lib/openai/models/beta/run_stream_event.rbi b/rbi/lib/openai/models/beta/run_stream_event.rbi index 2b6188fc..cedd2400 100644 --- a/rbi/lib/openai/models/beta/run_stream_event.rbi +++ b/rbi/lib/openai/models/beta/run_stream_event.rbi @@ -6,15 +6,15 @@ module OpenAI # Occurs when a new # [run](https://platform.openai.com/docs/api-reference/runs/object) is created. module RunStreamEvent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union - class ThreadRunCreated < OpenAI::BaseModel + class ThreadRunCreated < OpenAI::Internal::Type::BaseModel # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Run) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -23,7 +23,7 @@ module OpenAI # Occurs when a new # [run](https://platform.openai.com/docs/api-reference/runs/object) is created. sig do - params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash), event: Symbol) + params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.run.created") @@ -34,13 +34,13 @@ module OpenAI end end - class ThreadRunQueued < OpenAI::BaseModel + class ThreadRunQueued < OpenAI::Internal::Type::BaseModel # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Run) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -49,7 +49,7 @@ module OpenAI # Occurs when a [run](https://platform.openai.com/docs/api-reference/runs/object) # moves to a `queued` status. sig do - params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash), event: Symbol) + params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.run.queued") @@ -60,13 +60,13 @@ module OpenAI end end - class ThreadRunInProgress < OpenAI::BaseModel + class ThreadRunInProgress < OpenAI::Internal::Type::BaseModel # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Run) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -75,7 +75,7 @@ module OpenAI # Occurs when a [run](https://platform.openai.com/docs/api-reference/runs/object) # moves to an `in_progress` status. sig do - params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash), event: Symbol) + params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.run.in_progress") @@ -86,13 +86,13 @@ module OpenAI end end - class ThreadRunRequiresAction < OpenAI::BaseModel + class ThreadRunRequiresAction < OpenAI::Internal::Type::BaseModel # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Run) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -101,7 +101,7 @@ module OpenAI # Occurs when a [run](https://platform.openai.com/docs/api-reference/runs/object) # moves to a `requires_action` status. sig do - params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash), event: Symbol) + params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.run.requires_action") @@ -112,13 +112,13 @@ module OpenAI end end - class ThreadRunCompleted < OpenAI::BaseModel + class ThreadRunCompleted < OpenAI::Internal::Type::BaseModel # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Run) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -127,7 +127,7 @@ module OpenAI # Occurs when a [run](https://platform.openai.com/docs/api-reference/runs/object) # is completed. sig do - params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash), event: Symbol) + params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.run.completed") @@ -138,13 +138,13 @@ module OpenAI end end - class ThreadRunIncomplete < OpenAI::BaseModel + class ThreadRunIncomplete < OpenAI::Internal::Type::BaseModel # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Run) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -153,7 +153,7 @@ module OpenAI # Occurs when a [run](https://platform.openai.com/docs/api-reference/runs/object) # ends with status `incomplete`. sig do - params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash), event: Symbol) + params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.run.incomplete") @@ -164,13 +164,13 @@ module OpenAI end end - class ThreadRunFailed < OpenAI::BaseModel + class ThreadRunFailed < OpenAI::Internal::Type::BaseModel # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Run) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -179,7 +179,7 @@ module OpenAI # Occurs when a [run](https://platform.openai.com/docs/api-reference/runs/object) # fails. sig do - params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash), event: Symbol) + params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.run.failed") @@ -190,13 +190,13 @@ module OpenAI end end - class ThreadRunCancelling < OpenAI::BaseModel + class ThreadRunCancelling < OpenAI::Internal::Type::BaseModel # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Run) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -205,7 +205,7 @@ module OpenAI # Occurs when a [run](https://platform.openai.com/docs/api-reference/runs/object) # moves to a `cancelling` status. sig do - params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash), event: Symbol) + params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.run.cancelling") @@ -216,13 +216,13 @@ module OpenAI end end - class ThreadRunCancelled < OpenAI::BaseModel + class ThreadRunCancelled < OpenAI::Internal::Type::BaseModel # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Run) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -231,7 +231,7 @@ module OpenAI # Occurs when a [run](https://platform.openai.com/docs/api-reference/runs/object) # is cancelled. sig do - params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash), event: Symbol) + params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.run.cancelled") @@ -242,13 +242,13 @@ module OpenAI end end - class ThreadRunExpired < OpenAI::BaseModel + class ThreadRunExpired < OpenAI::Internal::Type::BaseModel # Represents an execution run on a # [thread](https://platform.openai.com/docs/api-reference/threads). sig { returns(OpenAI::Models::Beta::Threads::Run) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -257,7 +257,7 @@ module OpenAI # Occurs when a [run](https://platform.openai.com/docs/api-reference/runs/object) # expires. sig do - params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::Util::AnyHash), event: Symbol) + params(data: T.any(OpenAI::Models::Beta::Threads::Run, OpenAI::Internal::AnyHash), event: Symbol) .returns(T.attached_class) end def self.new(data:, event: :"thread.run.expired") diff --git a/rbi/lib/openai/models/beta/thread.rbi b/rbi/lib/openai/models/beta/thread.rbi index 052a7fc3..b6dd7fda 100644 --- a/rbi/lib/openai/models/beta/thread.rbi +++ b/rbi/lib/openai/models/beta/thread.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Beta - class Thread < OpenAI::BaseModel + class Thread < OpenAI::Internal::Type::BaseModel # The identifier, which can be referenced in API endpoints. sig { returns(String) } attr_accessor :id @@ -34,7 +34,7 @@ module OpenAI sig do params( - tool_resources: T.nilable(T.any(OpenAI::Models::Beta::Thread::ToolResources, OpenAI::Internal::Util::AnyHash)) + tool_resources: T.nilable(T.any(OpenAI::Models::Beta::Thread::ToolResources, OpenAI::Internal::AnyHash)) ) .void end @@ -47,7 +47,7 @@ module OpenAI id: String, created_at: Integer, metadata: T.nilable(T::Hash[Symbol, String]), - tool_resources: T.nilable(T.any(OpenAI::Models::Beta::Thread::ToolResources, OpenAI::Internal::Util::AnyHash)), + tool_resources: T.nilable(T.any(OpenAI::Models::Beta::Thread::ToolResources, OpenAI::Internal::AnyHash)), object: Symbol ) .returns(T.attached_class) @@ -70,13 +70,13 @@ module OpenAI def to_hash end - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel sig { returns(T.nilable(OpenAI::Models::Beta::Thread::ToolResources::CodeInterpreter)) } attr_reader :code_interpreter sig do params( - code_interpreter: T.any(OpenAI::Models::Beta::Thread::ToolResources::CodeInterpreter, OpenAI::Internal::Util::AnyHash) + code_interpreter: T.any(OpenAI::Models::Beta::Thread::ToolResources::CodeInterpreter, OpenAI::Internal::AnyHash) ) .void end @@ -87,7 +87,7 @@ module OpenAI sig do params( - file_search: T.any(OpenAI::Models::Beta::Thread::ToolResources::FileSearch, OpenAI::Internal::Util::AnyHash) + file_search: T.any(OpenAI::Models::Beta::Thread::ToolResources::FileSearch, OpenAI::Internal::AnyHash) ) .void end @@ -99,8 +99,8 @@ module OpenAI # tool requires a list of vector store IDs. sig do params( - code_interpreter: T.any(OpenAI::Models::Beta::Thread::ToolResources::CodeInterpreter, OpenAI::Internal::Util::AnyHash), - file_search: T.any(OpenAI::Models::Beta::Thread::ToolResources::FileSearch, OpenAI::Internal::Util::AnyHash) + code_interpreter: T.any(OpenAI::Models::Beta::Thread::ToolResources::CodeInterpreter, OpenAI::Internal::AnyHash), + file_search: T.any(OpenAI::Models::Beta::Thread::ToolResources::FileSearch, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -119,7 +119,7 @@ module OpenAI def to_hash end - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel # A list of [file](https://platform.openai.com/docs/api-reference/files) IDs made # available to the `code_interpreter` tool. There can be a maximum of 20 files # associated with the tool. @@ -138,7 +138,7 @@ module OpenAI end end - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # The # [vector store](https://platform.openai.com/docs/api-reference/vector-stores/object) # attached to this thread. There can be a maximum of 1 vector store attached to diff --git a/rbi/lib/openai/models/beta/thread_create_and_run_params.rbi b/rbi/lib/openai/models/beta/thread_create_and_run_params.rbi index c99ab223..c0aab016 100644 --- a/rbi/lib/openai/models/beta/thread_create_and_run_params.rbi +++ b/rbi/lib/openai/models/beta/thread_create_and_run_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Beta - class ThreadCreateAndRunParams < OpenAI::BaseModel + class ThreadCreateAndRunParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -105,9 +105,7 @@ module OpenAI attr_reader :thread sig do - params( - thread: T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread, OpenAI::Internal::Util::AnyHash) - ) + params(thread: T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread, OpenAI::Internal::AnyHash)) .void end attr_writer :thread @@ -140,9 +138,7 @@ module OpenAI sig do params( - tool_resources: T.nilable( - T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::ToolResources, OpenAI::Internal::Util::AnyHash) - ) + tool_resources: T.nilable(T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::ToolResources, OpenAI::Internal::AnyHash)) ) .void end @@ -181,7 +177,7 @@ module OpenAI sig do params( truncation_strategy: T.nilable( - T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::TruncationStrategy, OpenAI::Internal::Util::AnyHash) + T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::TruncationStrategy, OpenAI::Internal::AnyHash) ) ) .void @@ -201,28 +197,26 @@ module OpenAI T.any( Symbol, OpenAI::Models::ResponseFormatText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::ResponseFormatJSONObject, OpenAI::Models::ResponseFormatJSONSchema ) ), temperature: T.nilable(Float), - thread: T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread, OpenAI::Internal::Util::AnyHash), + thread: T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread, OpenAI::Internal::AnyHash), tool_choice: T.nilable( T.any( OpenAI::Models::Beta::AssistantToolChoiceOption::Auto::OrSymbol, OpenAI::Models::Beta::AssistantToolChoice, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ), - tool_resources: T.nilable( - T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::ToolResources, OpenAI::Internal::Util::AnyHash) - ), + tool_resources: T.nilable(T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::ToolResources, OpenAI::Internal::AnyHash)), tools: T.nilable( T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::FileSearchTool, OpenAI::Models::Beta::FunctionTool ) @@ -230,9 +224,9 @@ module OpenAI ), top_p: T.nilable(Float), truncation_strategy: T.nilable( - T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::TruncationStrategy, OpenAI::Internal::Util::AnyHash) + T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::TruncationStrategy, OpenAI::Internal::AnyHash) ), - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -307,14 +301,14 @@ module OpenAI # model associated with the assistant. If not, the model associated with the # assistant will be used. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, OpenAI::Models::ChatModel::OrSymbol]) } def self.variants end end - class Thread < OpenAI::BaseModel + class Thread < OpenAI::Internal::Type::BaseModel # A list of [messages](https://platform.openai.com/docs/api-reference/messages) to # start the thread with. sig { returns(T.nilable(T::Array[OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::Message])) } @@ -322,7 +316,7 @@ module OpenAI sig do params( - messages: T::Array[T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::Message, OpenAI::Internal::Util::AnyHash)] + messages: T::Array[T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::Message, OpenAI::Internal::AnyHash)] ) .void end @@ -347,10 +341,7 @@ module OpenAI sig do params( tool_resources: T.nilable( - T.any( - OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources, - OpenAI::Internal::Util::AnyHash - ) + T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources, OpenAI::Internal::AnyHash) ) ) .void @@ -361,13 +352,10 @@ module OpenAI # an empty thread will be created. sig do params( - messages: T::Array[T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::Message, OpenAI::Internal::Util::AnyHash)], + messages: T::Array[T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::Message, OpenAI::Internal::AnyHash)], metadata: T.nilable(T::Hash[Symbol, String]), tool_resources: T.nilable( - T.any( - OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources, - OpenAI::Internal::Util::AnyHash - ) + T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources, OpenAI::Internal::AnyHash) ) ) .returns(T.attached_class) @@ -388,7 +376,7 @@ module OpenAI def to_hash end - class Message < OpenAI::BaseModel + class Message < OpenAI::Internal::Type::BaseModel # The text contents of the message. sig do returns( @@ -435,7 +423,7 @@ module OpenAI T::Array[ T.any( OpenAI::Models::Beta::Threads::ImageFileContentBlock, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::ImageURLContentBlock, OpenAI::Models::Beta::Threads::TextContentBlockParam ) @@ -446,7 +434,7 @@ module OpenAI T::Array[ T.any( OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::Message::Attachment, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ] ), @@ -482,7 +470,7 @@ module OpenAI # The text contents of the message. module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override @@ -504,7 +492,7 @@ module OpenAI MessageContentPartParamArray = T.let( - OpenAI::ArrayOf[union: OpenAI::Models::Beta::Threads::MessageContentPartParam], + OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::Threads::MessageContentPartParam], OpenAI::Internal::Type::Converter ) end @@ -516,7 +504,7 @@ module OpenAI # - `assistant`: Indicates the message is generated by the assistant. Use this # value to insert messages from the assistant into the conversation. module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::Message::Role) } @@ -535,7 +523,7 @@ module OpenAI end end - class Attachment < OpenAI::BaseModel + class Attachment < OpenAI::Internal::Type::BaseModel # The ID of the file to attach to the message. sig { returns(T.nilable(String)) } attr_reader :file_id @@ -563,7 +551,7 @@ module OpenAI tools: T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::Message::Attachment::Tool::FileSearch ) ] @@ -578,7 +566,7 @@ module OpenAI tools: T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::Message::Attachment::Tool::FileSearch ) ] @@ -606,9 +594,9 @@ module OpenAI end module Tool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # The type of tool being defined: `file_search` sig { returns(Symbol) } attr_accessor :type @@ -634,7 +622,7 @@ module OpenAI end end - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel sig { returns(T.nilable(OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::CodeInterpreter)) } attr_reader :code_interpreter @@ -642,7 +630,7 @@ module OpenAI params( code_interpreter: T.any( OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::CodeInterpreter, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .void @@ -656,7 +644,7 @@ module OpenAI params( file_search: T.any( OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .void @@ -671,11 +659,11 @@ module OpenAI params( code_interpreter: T.any( OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::CodeInterpreter, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ), file_search: T.any( OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .returns(T.attached_class) @@ -695,7 +683,7 @@ module OpenAI def to_hash end - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel # A list of [file](https://platform.openai.com/docs/api-reference/files) IDs made # available to the `code_interpreter` tool. There can be a maximum of 20 files # associated with the tool. @@ -714,7 +702,7 @@ module OpenAI end end - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # The # [vector store](https://platform.openai.com/docs/api-reference/vector-stores/object) # attached to this thread. There can be a maximum of 1 vector store attached to @@ -743,7 +731,7 @@ module OpenAI vector_stores: T::Array[ T.any( OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch::VectorStore, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ] ) @@ -757,7 +745,7 @@ module OpenAI vector_stores: T::Array[ T.any( OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch::VectorStore, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ] ) @@ -778,7 +766,7 @@ module OpenAI def to_hash end - class VectorStore < OpenAI::BaseModel + class VectorStore < OpenAI::Internal::Type::BaseModel # The chunking strategy used to chunk the file(s). If not set, will use the `auto` # strategy. sig do @@ -797,7 +785,7 @@ module OpenAI params( chunking_strategy: T.any( OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Auto, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static ) ) @@ -827,7 +815,7 @@ module OpenAI params( chunking_strategy: T.any( OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Auto, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static ), file_ids: T::Array[String], @@ -857,9 +845,9 @@ module OpenAI # The chunking strategy used to chunk the file(s). If not set, will use the `auto` # strategy. module ChunkingStrategy - extend OpenAI::Union + extend OpenAI::Internal::Type::Union - class Auto < OpenAI::BaseModel + class Auto < OpenAI::Internal::Type::BaseModel # Always `auto`. sig { returns(Symbol) } attr_accessor :type @@ -875,7 +863,7 @@ module OpenAI end end - class Static < OpenAI::BaseModel + class Static < OpenAI::Internal::Type::BaseModel sig do returns( OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static::Static @@ -887,7 +875,7 @@ module OpenAI params( static: T.any( OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static::Static, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .void @@ -902,7 +890,7 @@ module OpenAI params( static: T.any( OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static::Static, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ), type: Symbol ) @@ -923,7 +911,7 @@ module OpenAI def to_hash end - class Static < OpenAI::BaseModel + class Static < OpenAI::Internal::Type::BaseModel # The number of tokens that overlap between chunks. The default value is `400`. # # Note that the overlap must not exceed half of `max_chunk_size_tokens`. @@ -966,7 +954,7 @@ module OpenAI end end - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel sig { returns(T.nilable(OpenAI::Models::Beta::ThreadCreateAndRunParams::ToolResources::CodeInterpreter)) } attr_reader :code_interpreter @@ -974,7 +962,7 @@ module OpenAI params( code_interpreter: T.any( OpenAI::Models::Beta::ThreadCreateAndRunParams::ToolResources::CodeInterpreter, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .void @@ -988,7 +976,7 @@ module OpenAI params( file_search: T.any( OpenAI::Models::Beta::ThreadCreateAndRunParams::ToolResources::FileSearch, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .void @@ -1003,11 +991,11 @@ module OpenAI params( code_interpreter: T.any( OpenAI::Models::Beta::ThreadCreateAndRunParams::ToolResources::CodeInterpreter, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ), file_search: T.any( OpenAI::Models::Beta::ThreadCreateAndRunParams::ToolResources::FileSearch, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .returns(T.attached_class) @@ -1027,7 +1015,7 @@ module OpenAI def to_hash end - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel # A list of [file](https://platform.openai.com/docs/api-reference/files) IDs made # available to the `code_interpreter` tool. There can be a maximum of 20 files # associated with the tool. @@ -1046,7 +1034,7 @@ module OpenAI end end - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # The ID of the # [vector store](https://platform.openai.com/docs/api-reference/vector-stores/object) # attached to this assistant. There can be a maximum of 1 vector store attached to @@ -1068,7 +1056,7 @@ module OpenAI end module Tool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override @@ -1080,7 +1068,7 @@ module OpenAI end end - class TruncationStrategy < OpenAI::BaseModel + class TruncationStrategy < OpenAI::Internal::Type::BaseModel # The truncation strategy to use for the thread. The default is `auto`. If set to # `last_messages`, the thread will be truncated to the n most recent messages in # the thread. When set to `auto`, messages in the middle of the thread will be @@ -1122,7 +1110,7 @@ module OpenAI # the thread. When set to `auto`, messages in the middle of the thread will be # dropped to fit the context length of the model, `max_prompt_tokens`. module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::ThreadCreateAndRunParams::TruncationStrategy::Type) } diff --git a/rbi/lib/openai/models/beta/thread_create_params.rbi b/rbi/lib/openai/models/beta/thread_create_params.rbi index 2e45ad3f..2f0b0063 100644 --- a/rbi/lib/openai/models/beta/thread_create_params.rbi +++ b/rbi/lib/openai/models/beta/thread_create_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Beta - class ThreadCreateParams < OpenAI::BaseModel + class ThreadCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -14,7 +14,7 @@ module OpenAI sig do params( - messages: T::Array[T.any(OpenAI::Models::Beta::ThreadCreateParams::Message, OpenAI::Internal::Util::AnyHash)] + messages: T::Array[T.any(OpenAI::Models::Beta::ThreadCreateParams::Message, OpenAI::Internal::AnyHash)] ) .void end @@ -38,7 +38,7 @@ module OpenAI sig do params( - tool_resources: T.nilable(T.any(OpenAI::Models::Beta::ThreadCreateParams::ToolResources, OpenAI::Internal::Util::AnyHash)) + tool_resources: T.nilable(T.any(OpenAI::Models::Beta::ThreadCreateParams::ToolResources, OpenAI::Internal::AnyHash)) ) .void end @@ -46,10 +46,10 @@ module OpenAI sig do params( - messages: T::Array[T.any(OpenAI::Models::Beta::ThreadCreateParams::Message, OpenAI::Internal::Util::AnyHash)], + messages: T::Array[T.any(OpenAI::Models::Beta::ThreadCreateParams::Message, OpenAI::Internal::AnyHash)], metadata: T.nilable(T::Hash[Symbol, String]), - tool_resources: T.nilable(T.any(OpenAI::Models::Beta::ThreadCreateParams::ToolResources, OpenAI::Internal::Util::AnyHash)), - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + tool_resources: T.nilable(T.any(OpenAI::Models::Beta::ThreadCreateParams::ToolResources, OpenAI::Internal::AnyHash)), + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -70,7 +70,7 @@ module OpenAI def to_hash end - class Message < OpenAI::BaseModel + class Message < OpenAI::Internal::Type::BaseModel # The text contents of the message. sig do returns( @@ -117,7 +117,7 @@ module OpenAI T::Array[ T.any( OpenAI::Models::Beta::Threads::ImageFileContentBlock, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::ImageURLContentBlock, OpenAI::Models::Beta::Threads::TextContentBlockParam ) @@ -125,7 +125,7 @@ module OpenAI ), role: OpenAI::Models::Beta::ThreadCreateParams::Message::Role::OrSymbol, attachments: T.nilable( - T::Array[T.any(OpenAI::Models::Beta::ThreadCreateParams::Message::Attachment, OpenAI::Internal::Util::AnyHash)] + T::Array[T.any(OpenAI::Models::Beta::ThreadCreateParams::Message::Attachment, OpenAI::Internal::AnyHash)] ), metadata: T.nilable(T::Hash[Symbol, String]) ) @@ -159,7 +159,7 @@ module OpenAI # The text contents of the message. module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override @@ -181,7 +181,7 @@ module OpenAI MessageContentPartParamArray = T.let( - OpenAI::ArrayOf[union: OpenAI::Models::Beta::Threads::MessageContentPartParam], + OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::Threads::MessageContentPartParam], OpenAI::Internal::Type::Converter ) end @@ -193,7 +193,7 @@ module OpenAI # - `assistant`: Indicates the message is generated by the assistant. Use this # value to insert messages from the assistant into the conversation. module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::ThreadCreateParams::Message::Role) } OrSymbol = @@ -207,7 +207,7 @@ module OpenAI end end - class Attachment < OpenAI::BaseModel + class Attachment < OpenAI::Internal::Type::BaseModel # The ID of the file to attach to the message. sig { returns(T.nilable(String)) } attr_reader :file_id @@ -235,7 +235,7 @@ module OpenAI tools: T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::ThreadCreateParams::Message::Attachment::Tool::FileSearch ) ] @@ -250,7 +250,7 @@ module OpenAI tools: T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::ThreadCreateParams::Message::Attachment::Tool::FileSearch ) ] @@ -278,9 +278,9 @@ module OpenAI end module Tool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # The type of tool being defined: `file_search` sig { returns(Symbol) } attr_accessor :type @@ -306,16 +306,13 @@ module OpenAI end end - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel sig { returns(T.nilable(OpenAI::Models::Beta::ThreadCreateParams::ToolResources::CodeInterpreter)) } attr_reader :code_interpreter sig do params( - code_interpreter: T.any( - OpenAI::Models::Beta::ThreadCreateParams::ToolResources::CodeInterpreter, - OpenAI::Internal::Util::AnyHash - ) + code_interpreter: T.any(OpenAI::Models::Beta::ThreadCreateParams::ToolResources::CodeInterpreter, OpenAI::Internal::AnyHash) ) .void end @@ -326,10 +323,7 @@ module OpenAI sig do params( - file_search: T.any( - OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch, - OpenAI::Internal::Util::AnyHash - ) + file_search: T.any(OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch, OpenAI::Internal::AnyHash) ) .void end @@ -341,14 +335,8 @@ module OpenAI # tool requires a list of vector store IDs. sig do params( - code_interpreter: T.any( - OpenAI::Models::Beta::ThreadCreateParams::ToolResources::CodeInterpreter, - OpenAI::Internal::Util::AnyHash - ), - file_search: T.any( - OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch, - OpenAI::Internal::Util::AnyHash - ) + code_interpreter: T.any(OpenAI::Models::Beta::ThreadCreateParams::ToolResources::CodeInterpreter, OpenAI::Internal::AnyHash), + file_search: T.any(OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -367,7 +355,7 @@ module OpenAI def to_hash end - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel # A list of [file](https://platform.openai.com/docs/api-reference/files) IDs made # available to the `code_interpreter` tool. There can be a maximum of 20 files # associated with the tool. @@ -386,7 +374,7 @@ module OpenAI end end - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # The # [vector store](https://platform.openai.com/docs/api-reference/vector-stores/object) # attached to this thread. There can be a maximum of 1 vector store attached to @@ -413,7 +401,7 @@ module OpenAI vector_stores: T::Array[ T.any( OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch::VectorStore, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ] ) @@ -427,7 +415,7 @@ module OpenAI vector_stores: T::Array[ T.any( OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch::VectorStore, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ] ) @@ -448,7 +436,7 @@ module OpenAI def to_hash end - class VectorStore < OpenAI::BaseModel + class VectorStore < OpenAI::Internal::Type::BaseModel # The chunking strategy used to chunk the file(s). If not set, will use the `auto` # strategy. sig do @@ -467,7 +455,7 @@ module OpenAI params( chunking_strategy: T.any( OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Auto, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static ) ) @@ -497,7 +485,7 @@ module OpenAI params( chunking_strategy: T.any( OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Auto, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static ), file_ids: T::Array[String], @@ -527,9 +515,9 @@ module OpenAI # The chunking strategy used to chunk the file(s). If not set, will use the `auto` # strategy. module ChunkingStrategy - extend OpenAI::Union + extend OpenAI::Internal::Type::Union - class Auto < OpenAI::BaseModel + class Auto < OpenAI::Internal::Type::BaseModel # Always `auto`. sig { returns(Symbol) } attr_accessor :type @@ -545,7 +533,7 @@ module OpenAI end end - class Static < OpenAI::BaseModel + class Static < OpenAI::Internal::Type::BaseModel sig do returns( OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static::Static @@ -557,7 +545,7 @@ module OpenAI params( static: T.any( OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static::Static, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .void @@ -572,7 +560,7 @@ module OpenAI params( static: T.any( OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static::Static, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ), type: Symbol ) @@ -593,7 +581,7 @@ module OpenAI def to_hash end - class Static < OpenAI::BaseModel + class Static < OpenAI::Internal::Type::BaseModel # The number of tokens that overlap between chunks. The default value is `400`. # # Note that the overlap must not exceed half of `max_chunk_size_tokens`. diff --git a/rbi/lib/openai/models/beta/thread_delete_params.rbi b/rbi/lib/openai/models/beta/thread_delete_params.rbi index 0a775418..be5184db 100644 --- a/rbi/lib/openai/models/beta/thread_delete_params.rbi +++ b/rbi/lib/openai/models/beta/thread_delete_params.rbi @@ -3,13 +3,17 @@ module OpenAI module Models module Beta - class ThreadDeleteParams < OpenAI::BaseModel + class ThreadDeleteParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters sig do - params(request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) - .returns(T.attached_class) + params( + request_options: T.any( + OpenAI::RequestOptions, + OpenAI::Internal::AnyHash + ) + ).returns(T.attached_class) end def self.new(request_options: {}) end diff --git a/rbi/lib/openai/models/beta/thread_deleted.rbi b/rbi/lib/openai/models/beta/thread_deleted.rbi index cc908f49..68939994 100644 --- a/rbi/lib/openai/models/beta/thread_deleted.rbi +++ b/rbi/lib/openai/models/beta/thread_deleted.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Beta - class ThreadDeleted < OpenAI::BaseModel + class ThreadDeleted < OpenAI::Internal::Type::BaseModel sig { returns(String) } attr_accessor :id diff --git a/rbi/lib/openai/models/beta/thread_retrieve_params.rbi b/rbi/lib/openai/models/beta/thread_retrieve_params.rbi index 502dbe33..e42b6057 100644 --- a/rbi/lib/openai/models/beta/thread_retrieve_params.rbi +++ b/rbi/lib/openai/models/beta/thread_retrieve_params.rbi @@ -3,13 +3,17 @@ module OpenAI module Models module Beta - class ThreadRetrieveParams < OpenAI::BaseModel + class ThreadRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters sig do - params(request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) - .returns(T.attached_class) + params( + request_options: T.any( + OpenAI::RequestOptions, + OpenAI::Internal::AnyHash + ) + ).returns(T.attached_class) end def self.new(request_options: {}) end diff --git a/rbi/lib/openai/models/beta/thread_stream_event.rbi b/rbi/lib/openai/models/beta/thread_stream_event.rbi index 34ddbb64..1f9f6531 100644 --- a/rbi/lib/openai/models/beta/thread_stream_event.rbi +++ b/rbi/lib/openai/models/beta/thread_stream_event.rbi @@ -3,13 +3,13 @@ module OpenAI module Models module Beta - class ThreadStreamEvent < OpenAI::BaseModel + class ThreadStreamEvent < OpenAI::Internal::Type::BaseModel # Represents a thread that contains # [messages](https://platform.openai.com/docs/api-reference/messages). sig { returns(OpenAI::Models::Beta::Thread) } attr_reader :data - sig { params(data: T.any(OpenAI::Models::Beta::Thread, OpenAI::Internal::Util::AnyHash)).void } + sig { params(data: T.any(OpenAI::Models::Beta::Thread, OpenAI::Internal::AnyHash)).void } attr_writer :data sig { returns(Symbol) } @@ -27,7 +27,7 @@ module OpenAI # created. sig do params( - data: T.any(OpenAI::Models::Beta::Thread, OpenAI::Internal::Util::AnyHash), + data: T.any(OpenAI::Models::Beta::Thread, OpenAI::Internal::AnyHash), enabled: T::Boolean, event: Symbol ) diff --git a/rbi/lib/openai/models/beta/thread_update_params.rbi b/rbi/lib/openai/models/beta/thread_update_params.rbi index 0f78f9cb..a1f2bf11 100644 --- a/rbi/lib/openai/models/beta/thread_update_params.rbi +++ b/rbi/lib/openai/models/beta/thread_update_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Beta - class ThreadUpdateParams < OpenAI::BaseModel + class ThreadUpdateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -25,7 +25,7 @@ module OpenAI sig do params( - tool_resources: T.nilable(T.any(OpenAI::Models::Beta::ThreadUpdateParams::ToolResources, OpenAI::Internal::Util::AnyHash)) + tool_resources: T.nilable(T.any(OpenAI::Models::Beta::ThreadUpdateParams::ToolResources, OpenAI::Internal::AnyHash)) ) .void end @@ -34,8 +34,8 @@ module OpenAI sig do params( metadata: T.nilable(T::Hash[Symbol, String]), - tool_resources: T.nilable(T.any(OpenAI::Models::Beta::ThreadUpdateParams::ToolResources, OpenAI::Internal::Util::AnyHash)), - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + tool_resources: T.nilable(T.any(OpenAI::Models::Beta::ThreadUpdateParams::ToolResources, OpenAI::Internal::AnyHash)), + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -55,16 +55,13 @@ module OpenAI def to_hash end - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel sig { returns(T.nilable(OpenAI::Models::Beta::ThreadUpdateParams::ToolResources::CodeInterpreter)) } attr_reader :code_interpreter sig do params( - code_interpreter: T.any( - OpenAI::Models::Beta::ThreadUpdateParams::ToolResources::CodeInterpreter, - OpenAI::Internal::Util::AnyHash - ) + code_interpreter: T.any(OpenAI::Models::Beta::ThreadUpdateParams::ToolResources::CodeInterpreter, OpenAI::Internal::AnyHash) ) .void end @@ -75,10 +72,7 @@ module OpenAI sig do params( - file_search: T.any( - OpenAI::Models::Beta::ThreadUpdateParams::ToolResources::FileSearch, - OpenAI::Internal::Util::AnyHash - ) + file_search: T.any(OpenAI::Models::Beta::ThreadUpdateParams::ToolResources::FileSearch, OpenAI::Internal::AnyHash) ) .void end @@ -90,14 +84,8 @@ module OpenAI # tool requires a list of vector store IDs. sig do params( - code_interpreter: T.any( - OpenAI::Models::Beta::ThreadUpdateParams::ToolResources::CodeInterpreter, - OpenAI::Internal::Util::AnyHash - ), - file_search: T.any( - OpenAI::Models::Beta::ThreadUpdateParams::ToolResources::FileSearch, - OpenAI::Internal::Util::AnyHash - ) + code_interpreter: T.any(OpenAI::Models::Beta::ThreadUpdateParams::ToolResources::CodeInterpreter, OpenAI::Internal::AnyHash), + file_search: T.any(OpenAI::Models::Beta::ThreadUpdateParams::ToolResources::FileSearch, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -116,7 +104,7 @@ module OpenAI def to_hash end - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel # A list of [file](https://platform.openai.com/docs/api-reference/files) IDs made # available to the `code_interpreter` tool. There can be a maximum of 20 files # associated with the tool. @@ -135,7 +123,7 @@ module OpenAI end end - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # The # [vector store](https://platform.openai.com/docs/api-reference/vector-stores/object) # attached to this thread. There can be a maximum of 1 vector store attached to diff --git a/rbi/lib/openai/models/beta/threads/annotation.rbi b/rbi/lib/openai/models/beta/threads/annotation.rbi index 02f8a33e..2579d047 100644 --- a/rbi/lib/openai/models/beta/threads/annotation.rbi +++ b/rbi/lib/openai/models/beta/threads/annotation.rbi @@ -8,7 +8,7 @@ module OpenAI # File associated with the assistant or the message. Generated when the assistant # uses the "file_search" tool to search files. module Annotation - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/beta/threads/annotation_delta.rbi b/rbi/lib/openai/models/beta/threads/annotation_delta.rbi index 133c007f..a0c01cad 100644 --- a/rbi/lib/openai/models/beta/threads/annotation_delta.rbi +++ b/rbi/lib/openai/models/beta/threads/annotation_delta.rbi @@ -8,7 +8,7 @@ module OpenAI # File associated with the assistant or the message. Generated when the assistant # uses the "file_search" tool to search files. module AnnotationDelta - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/beta/threads/file_citation_annotation.rbi b/rbi/lib/openai/models/beta/threads/file_citation_annotation.rbi index 2aaa8fe0..b87f7920 100644 --- a/rbi/lib/openai/models/beta/threads/file_citation_annotation.rbi +++ b/rbi/lib/openai/models/beta/threads/file_citation_annotation.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class FileCitationAnnotation < OpenAI::BaseModel + class FileCitationAnnotation < OpenAI::Internal::Type::BaseModel sig { returns(Integer) } attr_accessor :end_index @@ -13,10 +13,7 @@ module OpenAI sig do params( - file_citation: T.any( - OpenAI::Models::Beta::Threads::FileCitationAnnotation::FileCitation, - OpenAI::Internal::Util::AnyHash - ) + file_citation: T.any(OpenAI::Models::Beta::Threads::FileCitationAnnotation::FileCitation, OpenAI::Internal::AnyHash) ) .void end @@ -39,10 +36,7 @@ module OpenAI sig do params( end_index: Integer, - file_citation: T.any( - OpenAI::Models::Beta::Threads::FileCitationAnnotation::FileCitation, - OpenAI::Internal::Util::AnyHash - ), + file_citation: T.any(OpenAI::Models::Beta::Threads::FileCitationAnnotation::FileCitation, OpenAI::Internal::AnyHash), start_index: Integer, text: String, type: Symbol @@ -67,7 +61,7 @@ module OpenAI def to_hash end - class FileCitation < OpenAI::BaseModel + class FileCitation < OpenAI::Internal::Type::BaseModel # The ID of the specific File the citation is from. sig { returns(String) } attr_accessor :file_id diff --git a/rbi/lib/openai/models/beta/threads/file_citation_delta_annotation.rbi b/rbi/lib/openai/models/beta/threads/file_citation_delta_annotation.rbi index 9de4eb01..c8a41dac 100644 --- a/rbi/lib/openai/models/beta/threads/file_citation_delta_annotation.rbi +++ b/rbi/lib/openai/models/beta/threads/file_citation_delta_annotation.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class FileCitationDeltaAnnotation < OpenAI::BaseModel + class FileCitationDeltaAnnotation < OpenAI::Internal::Type::BaseModel # The index of the annotation in the text content part. sig { returns(Integer) } attr_accessor :index @@ -24,10 +24,7 @@ module OpenAI sig do params( - file_citation: T.any( - OpenAI::Models::Beta::Threads::FileCitationDeltaAnnotation::FileCitation, - OpenAI::Internal::Util::AnyHash - ) + file_citation: T.any(OpenAI::Models::Beta::Threads::FileCitationDeltaAnnotation::FileCitation, OpenAI::Internal::AnyHash) ) .void end @@ -53,10 +50,7 @@ module OpenAI params( index: Integer, end_index: Integer, - file_citation: T.any( - OpenAI::Models::Beta::Threads::FileCitationDeltaAnnotation::FileCitation, - OpenAI::Internal::Util::AnyHash - ), + file_citation: T.any(OpenAI::Models::Beta::Threads::FileCitationDeltaAnnotation::FileCitation, OpenAI::Internal::AnyHash), start_index: Integer, text: String, type: Symbol @@ -89,7 +83,7 @@ module OpenAI def to_hash end - class FileCitation < OpenAI::BaseModel + class FileCitation < OpenAI::Internal::Type::BaseModel # The ID of the specific File the citation is from. sig { returns(T.nilable(String)) } attr_reader :file_id diff --git a/rbi/lib/openai/models/beta/threads/file_path_annotation.rbi b/rbi/lib/openai/models/beta/threads/file_path_annotation.rbi index 7e8487a3..d3f54f98 100644 --- a/rbi/lib/openai/models/beta/threads/file_path_annotation.rbi +++ b/rbi/lib/openai/models/beta/threads/file_path_annotation.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class FilePathAnnotation < OpenAI::BaseModel + class FilePathAnnotation < OpenAI::Internal::Type::BaseModel sig { returns(Integer) } attr_accessor :end_index @@ -13,7 +13,7 @@ module OpenAI sig do params( - file_path: T.any(OpenAI::Models::Beta::Threads::FilePathAnnotation::FilePath, OpenAI::Internal::Util::AnyHash) + file_path: T.any(OpenAI::Models::Beta::Threads::FilePathAnnotation::FilePath, OpenAI::Internal::AnyHash) ) .void end @@ -35,7 +35,7 @@ module OpenAI sig do params( end_index: Integer, - file_path: T.any(OpenAI::Models::Beta::Threads::FilePathAnnotation::FilePath, OpenAI::Internal::Util::AnyHash), + file_path: T.any(OpenAI::Models::Beta::Threads::FilePathAnnotation::FilePath, OpenAI::Internal::AnyHash), start_index: Integer, text: String, type: Symbol @@ -60,7 +60,7 @@ module OpenAI def to_hash end - class FilePath < OpenAI::BaseModel + class FilePath < OpenAI::Internal::Type::BaseModel # The ID of the file that was generated. sig { returns(String) } attr_accessor :file_id diff --git a/rbi/lib/openai/models/beta/threads/file_path_delta_annotation.rbi b/rbi/lib/openai/models/beta/threads/file_path_delta_annotation.rbi index 51269c89..a47885b8 100644 --- a/rbi/lib/openai/models/beta/threads/file_path_delta_annotation.rbi +++ b/rbi/lib/openai/models/beta/threads/file_path_delta_annotation.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class FilePathDeltaAnnotation < OpenAI::BaseModel + class FilePathDeltaAnnotation < OpenAI::Internal::Type::BaseModel # The index of the annotation in the text content part. sig { returns(Integer) } attr_accessor :index @@ -24,7 +24,7 @@ module OpenAI sig do params( - file_path: T.any(OpenAI::Models::Beta::Threads::FilePathDeltaAnnotation::FilePath, OpenAI::Internal::Util::AnyHash) + file_path: T.any(OpenAI::Models::Beta::Threads::FilePathDeltaAnnotation::FilePath, OpenAI::Internal::AnyHash) ) .void end @@ -49,7 +49,7 @@ module OpenAI params( index: Integer, end_index: Integer, - file_path: T.any(OpenAI::Models::Beta::Threads::FilePathDeltaAnnotation::FilePath, OpenAI::Internal::Util::AnyHash), + file_path: T.any(OpenAI::Models::Beta::Threads::FilePathDeltaAnnotation::FilePath, OpenAI::Internal::AnyHash), start_index: Integer, text: String, type: Symbol @@ -75,7 +75,7 @@ module OpenAI def to_hash end - class FilePath < OpenAI::BaseModel + class FilePath < OpenAI::Internal::Type::BaseModel # The ID of the file that was generated. sig { returns(T.nilable(String)) } attr_reader :file_id diff --git a/rbi/lib/openai/models/beta/threads/image_file.rbi b/rbi/lib/openai/models/beta/threads/image_file.rbi index 2ce2e93b..f16895d6 100644 --- a/rbi/lib/openai/models/beta/threads/image_file.rbi +++ b/rbi/lib/openai/models/beta/threads/image_file.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class ImageFile < OpenAI::BaseModel + class ImageFile < OpenAI::Internal::Type::BaseModel # The [File](https://platform.openai.com/docs/api-reference/files) ID of the image # in the message content. Set `purpose="vision"` when uploading the File if you # need to later display the file content. @@ -33,7 +33,7 @@ module OpenAI # Specifies the detail level of the image if specified by the user. `low` uses # fewer tokens, you can opt in to high resolution using `high`. module Detail - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::ImageFile::Detail) } OrSymbol = diff --git a/rbi/lib/openai/models/beta/threads/image_file_content_block.rbi b/rbi/lib/openai/models/beta/threads/image_file_content_block.rbi index 301843bc..133cd72c 100644 --- a/rbi/lib/openai/models/beta/threads/image_file_content_block.rbi +++ b/rbi/lib/openai/models/beta/threads/image_file_content_block.rbi @@ -4,11 +4,11 @@ module OpenAI module Models module Beta module Threads - class ImageFileContentBlock < OpenAI::BaseModel + class ImageFileContentBlock < OpenAI::Internal::Type::BaseModel sig { returns(OpenAI::Models::Beta::Threads::ImageFile) } attr_reader :image_file - sig { params(image_file: T.any(OpenAI::Models::Beta::Threads::ImageFile, OpenAI::Internal::Util::AnyHash)).void } + sig { params(image_file: T.any(OpenAI::Models::Beta::Threads::ImageFile, OpenAI::Internal::AnyHash)).void } attr_writer :image_file # Always `image_file`. @@ -19,7 +19,7 @@ module OpenAI # in the content of a message. sig do params( - image_file: T.any(OpenAI::Models::Beta::Threads::ImageFile, OpenAI::Internal::Util::AnyHash), + image_file: T.any(OpenAI::Models::Beta::Threads::ImageFile, OpenAI::Internal::AnyHash), type: Symbol ) .returns(T.attached_class) diff --git a/rbi/lib/openai/models/beta/threads/image_file_delta.rbi b/rbi/lib/openai/models/beta/threads/image_file_delta.rbi index a0e96760..6d41ec50 100644 --- a/rbi/lib/openai/models/beta/threads/image_file_delta.rbi +++ b/rbi/lib/openai/models/beta/threads/image_file_delta.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class ImageFileDelta < OpenAI::BaseModel + class ImageFileDelta < OpenAI::Internal::Type::BaseModel # Specifies the detail level of the image if specified by the user. `low` uses # fewer tokens, you can opt in to high resolution using `high`. sig { returns(T.nilable(OpenAI::Models::Beta::Threads::ImageFileDelta::Detail::TaggedSymbol)) } @@ -39,7 +39,7 @@ module OpenAI # Specifies the detail level of the image if specified by the user. `low` uses # fewer tokens, you can opt in to high resolution using `high`. module Detail - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::ImageFileDelta::Detail) } OrSymbol = diff --git a/rbi/lib/openai/models/beta/threads/image_file_delta_block.rbi b/rbi/lib/openai/models/beta/threads/image_file_delta_block.rbi index 72d55cbb..7cb2f9a0 100644 --- a/rbi/lib/openai/models/beta/threads/image_file_delta_block.rbi +++ b/rbi/lib/openai/models/beta/threads/image_file_delta_block.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class ImageFileDeltaBlock < OpenAI::BaseModel + class ImageFileDeltaBlock < OpenAI::Internal::Type::BaseModel # The index of the content part in the message. sig { returns(Integer) } attr_accessor :index @@ -16,10 +16,7 @@ module OpenAI sig { returns(T.nilable(OpenAI::Models::Beta::Threads::ImageFileDelta)) } attr_reader :image_file - sig do - params(image_file: T.any(OpenAI::Models::Beta::Threads::ImageFileDelta, OpenAI::Internal::Util::AnyHash)) - .void - end + sig { params(image_file: T.any(OpenAI::Models::Beta::Threads::ImageFileDelta, OpenAI::Internal::AnyHash)).void } attr_writer :image_file # References an image [File](https://platform.openai.com/docs/api-reference/files) @@ -27,7 +24,7 @@ module OpenAI sig do params( index: Integer, - image_file: T.any(OpenAI::Models::Beta::Threads::ImageFileDelta, OpenAI::Internal::Util::AnyHash), + image_file: T.any(OpenAI::Models::Beta::Threads::ImageFileDelta, OpenAI::Internal::AnyHash), type: Symbol ) .returns(T.attached_class) diff --git a/rbi/lib/openai/models/beta/threads/image_url.rbi b/rbi/lib/openai/models/beta/threads/image_url.rbi index 0795a0a5..7bfaa664 100644 --- a/rbi/lib/openai/models/beta/threads/image_url.rbi +++ b/rbi/lib/openai/models/beta/threads/image_url.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class ImageURL < OpenAI::BaseModel + class ImageURL < OpenAI::Internal::Type::BaseModel # The external URL of the image, must be a supported image types: jpeg, jpg, png, # gif, webp. sig { returns(String) } @@ -32,7 +32,7 @@ module OpenAI # Specifies the detail level of the image. `low` uses fewer tokens, you can opt in # to high resolution using `high`. Default value is `auto` module Detail - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::ImageURL::Detail) } OrSymbol = diff --git a/rbi/lib/openai/models/beta/threads/image_url_content_block.rbi b/rbi/lib/openai/models/beta/threads/image_url_content_block.rbi index ec7592c9..2dff246b 100644 --- a/rbi/lib/openai/models/beta/threads/image_url_content_block.rbi +++ b/rbi/lib/openai/models/beta/threads/image_url_content_block.rbi @@ -4,11 +4,11 @@ module OpenAI module Models module Beta module Threads - class ImageURLContentBlock < OpenAI::BaseModel + class ImageURLContentBlock < OpenAI::Internal::Type::BaseModel sig { returns(OpenAI::Models::Beta::Threads::ImageURL) } attr_reader :image_url - sig { params(image_url: T.any(OpenAI::Models::Beta::Threads::ImageURL, OpenAI::Internal::Util::AnyHash)).void } + sig { params(image_url: T.any(OpenAI::Models::Beta::Threads::ImageURL, OpenAI::Internal::AnyHash)).void } attr_writer :image_url # The type of the content part. @@ -17,10 +17,7 @@ module OpenAI # References an image URL in the content of a message. sig do - params( - image_url: T.any(OpenAI::Models::Beta::Threads::ImageURL, OpenAI::Internal::Util::AnyHash), - type: Symbol - ) + params(image_url: T.any(OpenAI::Models::Beta::Threads::ImageURL, OpenAI::Internal::AnyHash), type: Symbol) .returns(T.attached_class) end def self.new(image_url:, type: :image_url) diff --git a/rbi/lib/openai/models/beta/threads/image_url_delta.rbi b/rbi/lib/openai/models/beta/threads/image_url_delta.rbi index aa4152eb..0fbc73d8 100644 --- a/rbi/lib/openai/models/beta/threads/image_url_delta.rbi +++ b/rbi/lib/openai/models/beta/threads/image_url_delta.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class ImageURLDelta < OpenAI::BaseModel + class ImageURLDelta < OpenAI::Internal::Type::BaseModel # Specifies the detail level of the image. `low` uses fewer tokens, you can opt in # to high resolution using `high`. sig { returns(T.nilable(OpenAI::Models::Beta::Threads::ImageURLDelta::Detail::TaggedSymbol)) } @@ -37,7 +37,7 @@ module OpenAI # Specifies the detail level of the image. `low` uses fewer tokens, you can opt in # to high resolution using `high`. module Detail - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::ImageURLDelta::Detail) } OrSymbol = diff --git a/rbi/lib/openai/models/beta/threads/image_url_delta_block.rbi b/rbi/lib/openai/models/beta/threads/image_url_delta_block.rbi index e337c120..75462fba 100644 --- a/rbi/lib/openai/models/beta/threads/image_url_delta_block.rbi +++ b/rbi/lib/openai/models/beta/threads/image_url_delta_block.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class ImageURLDeltaBlock < OpenAI::BaseModel + class ImageURLDeltaBlock < OpenAI::Internal::Type::BaseModel # The index of the content part in the message. sig { returns(Integer) } attr_accessor :index @@ -16,17 +16,14 @@ module OpenAI sig { returns(T.nilable(OpenAI::Models::Beta::Threads::ImageURLDelta)) } attr_reader :image_url - sig do - params(image_url: T.any(OpenAI::Models::Beta::Threads::ImageURLDelta, OpenAI::Internal::Util::AnyHash)) - .void - end + sig { params(image_url: T.any(OpenAI::Models::Beta::Threads::ImageURLDelta, OpenAI::Internal::AnyHash)).void } attr_writer :image_url # References an image URL in the content of a message. sig do params( index: Integer, - image_url: T.any(OpenAI::Models::Beta::Threads::ImageURLDelta, OpenAI::Internal::Util::AnyHash), + image_url: T.any(OpenAI::Models::Beta::Threads::ImageURLDelta, OpenAI::Internal::AnyHash), type: Symbol ) .returns(T.attached_class) diff --git a/rbi/lib/openai/models/beta/threads/message.rbi b/rbi/lib/openai/models/beta/threads/message.rbi index 03266cc5..6c7e5c86 100644 --- a/rbi/lib/openai/models/beta/threads/message.rbi +++ b/rbi/lib/openai/models/beta/threads/message.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class Message < OpenAI::BaseModel + class Message < OpenAI::Internal::Type::BaseModel # The identifier, which can be referenced in API endpoints. sig { returns(String) } attr_accessor :id @@ -52,9 +52,7 @@ module OpenAI sig do params( - incomplete_details: T.nilable( - T.any(OpenAI::Models::Beta::Threads::Message::IncompleteDetails, OpenAI::Internal::Util::AnyHash) - ) + incomplete_details: T.nilable(T.any(OpenAI::Models::Beta::Threads::Message::IncompleteDetails, OpenAI::Internal::AnyHash)) ) .void end @@ -99,14 +97,12 @@ module OpenAI params( id: String, assistant_id: T.nilable(String), - attachments: T.nilable( - T::Array[T.any(OpenAI::Models::Beta::Threads::Message::Attachment, OpenAI::Internal::Util::AnyHash)] - ), + attachments: T.nilable(T::Array[T.any(OpenAI::Models::Beta::Threads::Message::Attachment, OpenAI::Internal::AnyHash)]), completed_at: T.nilable(Integer), content: T::Array[ T.any( OpenAI::Models::Beta::Threads::ImageFileContentBlock, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::ImageURLContentBlock, OpenAI::Models::Beta::Threads::TextContentBlock, OpenAI::Models::Beta::Threads::RefusalContentBlock @@ -114,9 +110,7 @@ module OpenAI ], created_at: Integer, incomplete_at: T.nilable(Integer), - incomplete_details: T.nilable( - T.any(OpenAI::Models::Beta::Threads::Message::IncompleteDetails, OpenAI::Internal::Util::AnyHash) - ), + incomplete_details: T.nilable(T.any(OpenAI::Models::Beta::Threads::Message::IncompleteDetails, OpenAI::Internal::AnyHash)), metadata: T.nilable(T::Hash[Symbol, String]), role: OpenAI::Models::Beta::Threads::Message::Role::OrSymbol, run_id: T.nilable(String), @@ -175,7 +169,7 @@ module OpenAI def to_hash end - class Attachment < OpenAI::BaseModel + class Attachment < OpenAI::Internal::Type::BaseModel # The ID of the file to attach to the message. sig { returns(T.nilable(String)) } attr_reader :file_id @@ -203,7 +197,7 @@ module OpenAI tools: T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::Message::Attachment::Tool::AssistantToolsFileSearchTypeOnly ) ] @@ -218,7 +212,7 @@ module OpenAI tools: T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::Message::Attachment::Tool::AssistantToolsFileSearchTypeOnly ) ] @@ -246,9 +240,9 @@ module OpenAI end module Tool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union - class AssistantToolsFileSearchTypeOnly < OpenAI::BaseModel + class AssistantToolsFileSearchTypeOnly < OpenAI::Internal::Type::BaseModel # The type of tool being defined: `file_search` sig { returns(Symbol) } attr_accessor :type @@ -273,7 +267,7 @@ module OpenAI end end - class IncompleteDetails < OpenAI::BaseModel + class IncompleteDetails < OpenAI::Internal::Type::BaseModel # The reason the message is incomplete. sig { returns(OpenAI::Models::Beta::Threads::Message::IncompleteDetails::Reason::TaggedSymbol) } attr_accessor :reason @@ -294,7 +288,7 @@ module OpenAI # The reason the message is incomplete. module Reason - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::Message::IncompleteDetails::Reason) } @@ -322,7 +316,7 @@ module OpenAI # The entity that produced the message. One of `user` or `assistant`. module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::Message::Role) } OrSymbol = @@ -339,7 +333,7 @@ module OpenAI # The status of the message, which can be either `in_progress`, `incomplete`, or # `completed`. module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::Message::Status) } OrSymbol = diff --git a/rbi/lib/openai/models/beta/threads/message_content.rbi b/rbi/lib/openai/models/beta/threads/message_content.rbi index 3b799135..eae2a6d4 100644 --- a/rbi/lib/openai/models/beta/threads/message_content.rbi +++ b/rbi/lib/openai/models/beta/threads/message_content.rbi @@ -7,7 +7,7 @@ module OpenAI # References an image [File](https://platform.openai.com/docs/api-reference/files) # in the content of a message. module MessageContent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/beta/threads/message_content_delta.rbi b/rbi/lib/openai/models/beta/threads/message_content_delta.rbi index 7cfeab9e..7be07281 100644 --- a/rbi/lib/openai/models/beta/threads/message_content_delta.rbi +++ b/rbi/lib/openai/models/beta/threads/message_content_delta.rbi @@ -7,7 +7,7 @@ module OpenAI # References an image [File](https://platform.openai.com/docs/api-reference/files) # in the content of a message. module MessageContentDelta - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/beta/threads/message_content_part_param.rbi b/rbi/lib/openai/models/beta/threads/message_content_part_param.rbi index e2e3c952..a6e0c71f 100644 --- a/rbi/lib/openai/models/beta/threads/message_content_part_param.rbi +++ b/rbi/lib/openai/models/beta/threads/message_content_part_param.rbi @@ -7,7 +7,7 @@ module OpenAI # References an image [File](https://platform.openai.com/docs/api-reference/files) # in the content of a message. module MessageContentPartParam - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/beta/threads/message_create_params.rbi b/rbi/lib/openai/models/beta/threads/message_create_params.rbi index 9ad49558..5ed1214c 100644 --- a/rbi/lib/openai/models/beta/threads/message_create_params.rbi +++ b/rbi/lib/openai/models/beta/threads/message_create_params.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class MessageCreateParams < OpenAI::BaseModel + class MessageCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -54,7 +54,7 @@ module OpenAI T::Array[ T.any( OpenAI::Models::Beta::Threads::ImageFileContentBlock, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::ImageURLContentBlock, OpenAI::Models::Beta::Threads::TextContentBlockParam ) @@ -62,10 +62,10 @@ module OpenAI ), role: OpenAI::Models::Beta::Threads::MessageCreateParams::Role::OrSymbol, attachments: T.nilable( - T::Array[T.any(OpenAI::Models::Beta::Threads::MessageCreateParams::Attachment, OpenAI::Internal::Util::AnyHash)] + T::Array[T.any(OpenAI::Models::Beta::Threads::MessageCreateParams::Attachment, OpenAI::Internal::AnyHash)] ), metadata: T.nilable(T::Hash[Symbol, String]), - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -98,7 +98,7 @@ module OpenAI # The text contents of the message. module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override @@ -120,7 +120,7 @@ module OpenAI MessageContentPartParamArray = T.let( - OpenAI::ArrayOf[union: OpenAI::Models::Beta::Threads::MessageContentPartParam], + OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::Threads::MessageContentPartParam], OpenAI::Internal::Type::Converter ) end @@ -132,7 +132,7 @@ module OpenAI # - `assistant`: Indicates the message is generated by the assistant. Use this # value to insert messages from the assistant into the conversation. module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::MessageCreateParams::Role) } OrSymbol = @@ -146,7 +146,7 @@ module OpenAI end end - class Attachment < OpenAI::BaseModel + class Attachment < OpenAI::Internal::Type::BaseModel # The ID of the file to attach to the message. sig { returns(T.nilable(String)) } attr_reader :file_id @@ -174,7 +174,7 @@ module OpenAI tools: T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::MessageCreateParams::Attachment::Tool::FileSearch ) ] @@ -189,7 +189,7 @@ module OpenAI tools: T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::MessageCreateParams::Attachment::Tool::FileSearch ) ] @@ -217,9 +217,9 @@ module OpenAI end module Tool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # The type of tool being defined: `file_search` sig { returns(Symbol) } attr_accessor :type diff --git a/rbi/lib/openai/models/beta/threads/message_delete_params.rbi b/rbi/lib/openai/models/beta/threads/message_delete_params.rbi index 71feebb1..0dfa827d 100644 --- a/rbi/lib/openai/models/beta/threads/message_delete_params.rbi +++ b/rbi/lib/openai/models/beta/threads/message_delete_params.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class MessageDeleteParams < OpenAI::BaseModel + class MessageDeleteParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -16,7 +16,7 @@ module OpenAI thread_id: String, request_options: T.any( OpenAI::RequestOptions, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .returns(T.attached_class) diff --git a/rbi/lib/openai/models/beta/threads/message_deleted.rbi b/rbi/lib/openai/models/beta/threads/message_deleted.rbi index df1dc2e2..fd6cfd06 100644 --- a/rbi/lib/openai/models/beta/threads/message_deleted.rbi +++ b/rbi/lib/openai/models/beta/threads/message_deleted.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class MessageDeleted < OpenAI::BaseModel + class MessageDeleted < OpenAI::Internal::Type::BaseModel sig { returns(String) } attr_accessor :id diff --git a/rbi/lib/openai/models/beta/threads/message_delta.rbi b/rbi/lib/openai/models/beta/threads/message_delta.rbi index 151f4efe..5ae1fdd1 100644 --- a/rbi/lib/openai/models/beta/threads/message_delta.rbi +++ b/rbi/lib/openai/models/beta/threads/message_delta.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class MessageDelta < OpenAI::BaseModel + class MessageDelta < OpenAI::Internal::Type::BaseModel # The content of the message in array of text and/or images. sig do returns( @@ -27,7 +27,7 @@ module OpenAI content: T::Array[ T.any( OpenAI::Models::Beta::Threads::ImageFileDeltaBlock, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::TextDeltaBlock, OpenAI::Models::Beta::Threads::RefusalDeltaBlock, OpenAI::Models::Beta::Threads::ImageURLDeltaBlock @@ -51,7 +51,7 @@ module OpenAI content: T::Array[ T.any( OpenAI::Models::Beta::Threads::ImageFileDeltaBlock, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::TextDeltaBlock, OpenAI::Models::Beta::Threads::RefusalDeltaBlock, OpenAI::Models::Beta::Threads::ImageURLDeltaBlock @@ -85,7 +85,7 @@ module OpenAI # The entity that produced the message. One of `user` or `assistant`. module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::MessageDelta::Role) } OrSymbol = diff --git a/rbi/lib/openai/models/beta/threads/message_delta_event.rbi b/rbi/lib/openai/models/beta/threads/message_delta_event.rbi index bae39ef8..3fd98304 100644 --- a/rbi/lib/openai/models/beta/threads/message_delta_event.rbi +++ b/rbi/lib/openai/models/beta/threads/message_delta_event.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class MessageDeltaEvent < OpenAI::BaseModel + class MessageDeltaEvent < OpenAI::Internal::Type::BaseModel # The identifier of the message, which can be referenced in API endpoints. sig { returns(String) } attr_accessor :id @@ -13,7 +13,7 @@ module OpenAI sig { returns(OpenAI::Models::Beta::Threads::MessageDelta) } attr_reader :delta - sig { params(delta: T.any(OpenAI::Models::Beta::Threads::MessageDelta, OpenAI::Internal::Util::AnyHash)).void } + sig { params(delta: T.any(OpenAI::Models::Beta::Threads::MessageDelta, OpenAI::Internal::AnyHash)).void } attr_writer :delta # The object type, which is always `thread.message.delta`. @@ -25,7 +25,7 @@ module OpenAI sig do params( id: String, - delta: T.any(OpenAI::Models::Beta::Threads::MessageDelta, OpenAI::Internal::Util::AnyHash), + delta: T.any(OpenAI::Models::Beta::Threads::MessageDelta, OpenAI::Internal::AnyHash), object: Symbol ) .returns(T.attached_class) diff --git a/rbi/lib/openai/models/beta/threads/message_list_params.rbi b/rbi/lib/openai/models/beta/threads/message_list_params.rbi index 3664eaeb..2a25adec 100644 --- a/rbi/lib/openai/models/beta/threads/message_list_params.rbi +++ b/rbi/lib/openai/models/beta/threads/message_list_params.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class MessageListParams < OpenAI::BaseModel + class MessageListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -58,7 +58,7 @@ module OpenAI limit: Integer, order: OpenAI::Models::Beta::Threads::MessageListParams::Order::OrSymbol, run_id: String, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -84,7 +84,7 @@ module OpenAI # Sort order by the `created_at` timestamp of the objects. `asc` for ascending # order and `desc` for descending order. module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::MessageListParams::Order) } OrSymbol = diff --git a/rbi/lib/openai/models/beta/threads/message_retrieve_params.rbi b/rbi/lib/openai/models/beta/threads/message_retrieve_params.rbi index bc78f6e0..62cc19a2 100644 --- a/rbi/lib/openai/models/beta/threads/message_retrieve_params.rbi +++ b/rbi/lib/openai/models/beta/threads/message_retrieve_params.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class MessageRetrieveParams < OpenAI::BaseModel + class MessageRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -16,7 +16,7 @@ module OpenAI thread_id: String, request_options: T.any( OpenAI::RequestOptions, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .returns(T.attached_class) diff --git a/rbi/lib/openai/models/beta/threads/message_update_params.rbi b/rbi/lib/openai/models/beta/threads/message_update_params.rbi index 0d2c98de..51074aac 100644 --- a/rbi/lib/openai/models/beta/threads/message_update_params.rbi +++ b/rbi/lib/openai/models/beta/threads/message_update_params.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class MessageUpdateParams < OpenAI::BaseModel + class MessageUpdateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -24,7 +24,7 @@ module OpenAI params( thread_id: String, metadata: T.nilable(T::Hash[Symbol, String]), - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end diff --git a/rbi/lib/openai/models/beta/threads/refusal_content_block.rbi b/rbi/lib/openai/models/beta/threads/refusal_content_block.rbi index b9350051..334ce796 100644 --- a/rbi/lib/openai/models/beta/threads/refusal_content_block.rbi +++ b/rbi/lib/openai/models/beta/threads/refusal_content_block.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class RefusalContentBlock < OpenAI::BaseModel + class RefusalContentBlock < OpenAI::Internal::Type::BaseModel sig { returns(String) } attr_accessor :refusal diff --git a/rbi/lib/openai/models/beta/threads/refusal_delta_block.rbi b/rbi/lib/openai/models/beta/threads/refusal_delta_block.rbi index 72b6092e..2caef1c7 100644 --- a/rbi/lib/openai/models/beta/threads/refusal_delta_block.rbi +++ b/rbi/lib/openai/models/beta/threads/refusal_delta_block.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class RefusalDeltaBlock < OpenAI::BaseModel + class RefusalDeltaBlock < OpenAI::Internal::Type::BaseModel # The index of the refusal part in the message. sig { returns(Integer) } attr_accessor :index diff --git a/rbi/lib/openai/models/beta/threads/required_action_function_tool_call.rbi b/rbi/lib/openai/models/beta/threads/required_action_function_tool_call.rbi index 762191b7..795f0789 100644 --- a/rbi/lib/openai/models/beta/threads/required_action_function_tool_call.rbi +++ b/rbi/lib/openai/models/beta/threads/required_action_function_tool_call.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class RequiredActionFunctionToolCall < OpenAI::BaseModel + class RequiredActionFunctionToolCall < OpenAI::Internal::Type::BaseModel # The ID of the tool call. This ID must be referenced when you submit the tool # outputs in using the # [Submit tool outputs to run](https://platform.openai.com/docs/api-reference/runs/submitToolOutputs) @@ -18,10 +18,7 @@ module OpenAI sig do params( - function: T.any( - OpenAI::Models::Beta::Threads::RequiredActionFunctionToolCall::Function, - OpenAI::Internal::Util::AnyHash - ) + function: T.any(OpenAI::Models::Beta::Threads::RequiredActionFunctionToolCall::Function, OpenAI::Internal::AnyHash) ) .void end @@ -36,10 +33,7 @@ module OpenAI sig do params( id: String, - function: T.any( - OpenAI::Models::Beta::Threads::RequiredActionFunctionToolCall::Function, - OpenAI::Internal::Util::AnyHash - ), + function: T.any(OpenAI::Models::Beta::Threads::RequiredActionFunctionToolCall::Function, OpenAI::Internal::AnyHash), type: Symbol ) .returns(T.attached_class) @@ -56,7 +50,7 @@ module OpenAI def to_hash end - class Function < OpenAI::BaseModel + class Function < OpenAI::Internal::Type::BaseModel # The arguments that the model expects you to pass to the function. sig { returns(String) } attr_accessor :arguments diff --git a/rbi/lib/openai/models/beta/threads/run.rbi b/rbi/lib/openai/models/beta/threads/run.rbi index d19d330f..47341501 100644 --- a/rbi/lib/openai/models/beta/threads/run.rbi +++ b/rbi/lib/openai/models/beta/threads/run.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class Run < OpenAI::BaseModel + class Run < OpenAI::Internal::Type::BaseModel # The identifier, which can be referenced in API endpoints. sig { returns(String) } attr_accessor :id @@ -42,7 +42,7 @@ module OpenAI sig do params( - incomplete_details: T.nilable(T.any(OpenAI::Models::Beta::Threads::Run::IncompleteDetails, OpenAI::Internal::Util::AnyHash)) + incomplete_details: T.nilable(T.any(OpenAI::Models::Beta::Threads::Run::IncompleteDetails, OpenAI::Internal::AnyHash)) ) .void end @@ -60,7 +60,7 @@ module OpenAI sig do params( - last_error: T.nilable(T.any(OpenAI::Models::Beta::Threads::Run::LastError, OpenAI::Internal::Util::AnyHash)) + last_error: T.nilable(T.any(OpenAI::Models::Beta::Threads::Run::LastError, OpenAI::Internal::AnyHash)) ) .void end @@ -108,7 +108,7 @@ module OpenAI sig do params( - required_action: T.nilable(T.any(OpenAI::Models::Beta::Threads::Run::RequiredAction, OpenAI::Internal::Util::AnyHash)) + required_action: T.nilable(T.any(OpenAI::Models::Beta::Threads::Run::RequiredAction, OpenAI::Internal::AnyHash)) ) .void end @@ -205,7 +205,7 @@ module OpenAI sig do params( - truncation_strategy: T.nilable(T.any(OpenAI::Models::Beta::Threads::Run::TruncationStrategy, OpenAI::Internal::Util::AnyHash)) + truncation_strategy: T.nilable(T.any(OpenAI::Models::Beta::Threads::Run::TruncationStrategy, OpenAI::Internal::AnyHash)) ) .void end @@ -216,12 +216,7 @@ module OpenAI sig { returns(T.nilable(OpenAI::Models::Beta::Threads::Run::Usage)) } attr_reader :usage - sig do - params( - usage: T.nilable(T.any(OpenAI::Models::Beta::Threads::Run::Usage, OpenAI::Internal::Util::AnyHash)) - ) - .void - end + sig { params(usage: T.nilable(T.any(OpenAI::Models::Beta::Threads::Run::Usage, OpenAI::Internal::AnyHash))).void } attr_writer :usage # The sampling temperature used for this run. If not set, defaults to 1. @@ -243,20 +238,20 @@ module OpenAI created_at: Integer, expires_at: T.nilable(Integer), failed_at: T.nilable(Integer), - incomplete_details: T.nilable(T.any(OpenAI::Models::Beta::Threads::Run::IncompleteDetails, OpenAI::Internal::Util::AnyHash)), + incomplete_details: T.nilable(T.any(OpenAI::Models::Beta::Threads::Run::IncompleteDetails, OpenAI::Internal::AnyHash)), instructions: String, - last_error: T.nilable(T.any(OpenAI::Models::Beta::Threads::Run::LastError, OpenAI::Internal::Util::AnyHash)), + last_error: T.nilable(T.any(OpenAI::Models::Beta::Threads::Run::LastError, OpenAI::Internal::AnyHash)), max_completion_tokens: T.nilable(Integer), max_prompt_tokens: T.nilable(Integer), metadata: T.nilable(T::Hash[Symbol, String]), model: String, parallel_tool_calls: T::Boolean, - required_action: T.nilable(T.any(OpenAI::Models::Beta::Threads::Run::RequiredAction, OpenAI::Internal::Util::AnyHash)), + required_action: T.nilable(T.any(OpenAI::Models::Beta::Threads::Run::RequiredAction, OpenAI::Internal::AnyHash)), response_format: T.nilable( T.any( Symbol, OpenAI::Models::ResponseFormatText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::ResponseFormatJSONObject, OpenAI::Models::ResponseFormatJSONSchema ) @@ -268,19 +263,19 @@ module OpenAI T.any( OpenAI::Models::Beta::AssistantToolChoiceOption::Auto::OrSymbol, OpenAI::Models::Beta::AssistantToolChoice, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ), tools: T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::FileSearchTool, OpenAI::Models::Beta::FunctionTool ) ], - truncation_strategy: T.nilable(T.any(OpenAI::Models::Beta::Threads::Run::TruncationStrategy, OpenAI::Internal::Util::AnyHash)), - usage: T.nilable(T.any(OpenAI::Models::Beta::Threads::Run::Usage, OpenAI::Internal::Util::AnyHash)), + truncation_strategy: T.nilable(T.any(OpenAI::Models::Beta::Threads::Run::TruncationStrategy, OpenAI::Internal::AnyHash)), + usage: T.nilable(T.any(OpenAI::Models::Beta::Threads::Run::Usage, OpenAI::Internal::AnyHash)), temperature: T.nilable(Float), top_p: T.nilable(Float), object: Symbol @@ -373,7 +368,7 @@ module OpenAI def to_hash end - class IncompleteDetails < OpenAI::BaseModel + class IncompleteDetails < OpenAI::Internal::Type::BaseModel # The reason why the run is incomplete. This will point to which specific token # limit was reached over the course of the run. sig { returns(T.nilable(OpenAI::Models::Beta::Threads::Run::IncompleteDetails::Reason::TaggedSymbol)) } @@ -398,7 +393,7 @@ module OpenAI # The reason why the run is incomplete. This will point to which specific token # limit was reached over the course of the run. module Reason - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::Run::IncompleteDetails::Reason) } @@ -416,7 +411,7 @@ module OpenAI end end - class LastError < OpenAI::BaseModel + class LastError < OpenAI::Internal::Type::BaseModel # One of `server_error`, `rate_limit_exceeded`, or `invalid_prompt`. sig { returns(OpenAI::Models::Beta::Threads::Run::LastError::Code::TaggedSymbol) } attr_accessor :code @@ -442,7 +437,7 @@ module OpenAI # One of `server_error`, `rate_limit_exceeded`, or `invalid_prompt`. module Code - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::Run::LastError::Code) } OrSymbol = @@ -460,17 +455,14 @@ module OpenAI end end - class RequiredAction < OpenAI::BaseModel + class RequiredAction < OpenAI::Internal::Type::BaseModel # Details on the tool outputs needed for this run to continue. sig { returns(OpenAI::Models::Beta::Threads::Run::RequiredAction::SubmitToolOutputs) } attr_reader :submit_tool_outputs sig do params( - submit_tool_outputs: T.any( - OpenAI::Models::Beta::Threads::Run::RequiredAction::SubmitToolOutputs, - OpenAI::Internal::Util::AnyHash - ) + submit_tool_outputs: T.any(OpenAI::Models::Beta::Threads::Run::RequiredAction::SubmitToolOutputs, OpenAI::Internal::AnyHash) ) .void end @@ -484,10 +476,7 @@ module OpenAI # is required. sig do params( - submit_tool_outputs: T.any( - OpenAI::Models::Beta::Threads::Run::RequiredAction::SubmitToolOutputs, - OpenAI::Internal::Util::AnyHash - ), + submit_tool_outputs: T.any(OpenAI::Models::Beta::Threads::Run::RequiredAction::SubmitToolOutputs, OpenAI::Internal::AnyHash), type: Symbol ) .returns(T.attached_class) @@ -504,7 +493,7 @@ module OpenAI def to_hash end - class SubmitToolOutputs < OpenAI::BaseModel + class SubmitToolOutputs < OpenAI::Internal::Type::BaseModel # A list of the relevant tool calls. sig { returns(T::Array[OpenAI::Models::Beta::Threads::RequiredActionFunctionToolCall]) } attr_accessor :tool_calls @@ -512,7 +501,7 @@ module OpenAI # Details on the tool outputs needed for this run to continue. sig do params( - tool_calls: T::Array[T.any(OpenAI::Models::Beta::Threads::RequiredActionFunctionToolCall, OpenAI::Internal::Util::AnyHash)] + tool_calls: T::Array[T.any(OpenAI::Models::Beta::Threads::RequiredActionFunctionToolCall, OpenAI::Internal::AnyHash)] ) .returns(T.attached_class) end @@ -525,7 +514,7 @@ module OpenAI end end - class TruncationStrategy < OpenAI::BaseModel + class TruncationStrategy < OpenAI::Internal::Type::BaseModel # The truncation strategy to use for the thread. The default is `auto`. If set to # `last_messages`, the thread will be truncated to the n most recent messages in # the thread. When set to `auto`, messages in the middle of the thread will be @@ -567,7 +556,7 @@ module OpenAI # the thread. When set to `auto`, messages in the middle of the thread will be # dropped to fit the context length of the model, `max_prompt_tokens`. module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::Run::TruncationStrategy::Type) } @@ -584,7 +573,7 @@ module OpenAI end end - class Usage < OpenAI::BaseModel + class Usage < OpenAI::Internal::Type::BaseModel # Number of completion tokens used over the course of the run. sig { returns(Integer) } attr_accessor :completion_tokens diff --git a/rbi/lib/openai/models/beta/threads/run_cancel_params.rbi b/rbi/lib/openai/models/beta/threads/run_cancel_params.rbi index 92c6ca3b..d82ac9df 100644 --- a/rbi/lib/openai/models/beta/threads/run_cancel_params.rbi +++ b/rbi/lib/openai/models/beta/threads/run_cancel_params.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class RunCancelParams < OpenAI::BaseModel + class RunCancelParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -16,7 +16,7 @@ module OpenAI thread_id: String, request_options: T.any( OpenAI::RequestOptions, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .returns(T.attached_class) diff --git a/rbi/lib/openai/models/beta/threads/run_create_params.rbi b/rbi/lib/openai/models/beta/threads/run_create_params.rbi index e7854e85..a34ebb8a 100644 --- a/rbi/lib/openai/models/beta/threads/run_create_params.rbi +++ b/rbi/lib/openai/models/beta/threads/run_create_params.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class RunCreateParams < OpenAI::BaseModel + class RunCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -185,7 +185,7 @@ module OpenAI sig do params( truncation_strategy: T.nilable( - T.any(OpenAI::Models::Beta::Threads::RunCreateParams::TruncationStrategy, OpenAI::Internal::Util::AnyHash) + T.any(OpenAI::Models::Beta::Threads::RunCreateParams::TruncationStrategy, OpenAI::Internal::AnyHash) ) ) .void @@ -198,7 +198,7 @@ module OpenAI include: T::Array[OpenAI::Models::Beta::Threads::Runs::RunStepInclude::OrSymbol], additional_instructions: T.nilable(String), additional_messages: T.nilable( - T::Array[T.any(OpenAI::Models::Beta::Threads::RunCreateParams::AdditionalMessage, OpenAI::Internal::Util::AnyHash)] + T::Array[T.any(OpenAI::Models::Beta::Threads::RunCreateParams::AdditionalMessage, OpenAI::Internal::AnyHash)] ), instructions: T.nilable(String), max_completion_tokens: T.nilable(Integer), @@ -211,7 +211,7 @@ module OpenAI T.any( Symbol, OpenAI::Models::ResponseFormatText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::ResponseFormatJSONObject, OpenAI::Models::ResponseFormatJSONSchema ) @@ -221,14 +221,14 @@ module OpenAI T.any( OpenAI::Models::Beta::AssistantToolChoiceOption::Auto::OrSymbol, OpenAI::Models::Beta::AssistantToolChoice, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ), tools: T.nilable( T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::FileSearchTool, OpenAI::Models::Beta::FunctionTool ) @@ -236,9 +236,9 @@ module OpenAI ), top_p: T.nilable(Float), truncation_strategy: T.nilable( - T.any(OpenAI::Models::Beta::Threads::RunCreateParams::TruncationStrategy, OpenAI::Internal::Util::AnyHash) + T.any(OpenAI::Models::Beta::Threads::RunCreateParams::TruncationStrategy, OpenAI::Internal::AnyHash) ), - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -312,7 +312,7 @@ module OpenAI def to_hash end - class AdditionalMessage < OpenAI::BaseModel + class AdditionalMessage < OpenAI::Internal::Type::BaseModel # The text contents of the message. sig do returns( @@ -363,7 +363,7 @@ module OpenAI T::Array[ T.any( OpenAI::Models::Beta::Threads::ImageFileContentBlock, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::ImageURLContentBlock, OpenAI::Models::Beta::Threads::TextContentBlockParam ) @@ -374,7 +374,7 @@ module OpenAI T::Array[ T.any( OpenAI::Models::Beta::Threads::RunCreateParams::AdditionalMessage::Attachment, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ] ), @@ -410,7 +410,7 @@ module OpenAI # The text contents of the message. module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override @@ -432,7 +432,7 @@ module OpenAI MessageContentPartParamArray = T.let( - OpenAI::ArrayOf[union: OpenAI::Models::Beta::Threads::MessageContentPartParam], + OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::Threads::MessageContentPartParam], OpenAI::Internal::Type::Converter ) end @@ -444,7 +444,7 @@ module OpenAI # - `assistant`: Indicates the message is generated by the assistant. Use this # value to insert messages from the assistant into the conversation. module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::RunCreateParams::AdditionalMessage::Role) } @@ -470,7 +470,7 @@ module OpenAI end end - class Attachment < OpenAI::BaseModel + class Attachment < OpenAI::Internal::Type::BaseModel # The ID of the file to attach to the message. sig { returns(T.nilable(String)) } attr_reader :file_id @@ -498,7 +498,7 @@ module OpenAI tools: T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::RunCreateParams::AdditionalMessage::Attachment::Tool::FileSearch ) ] @@ -513,7 +513,7 @@ module OpenAI tools: T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::RunCreateParams::AdditionalMessage::Attachment::Tool::FileSearch ) ] @@ -541,9 +541,9 @@ module OpenAI end module Tool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # The type of tool being defined: `file_search` sig { returns(Symbol) } attr_accessor :type @@ -574,14 +574,14 @@ module OpenAI # model associated with the assistant. If not, the model associated with the # assistant will be used. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, OpenAI::Models::ChatModel::OrSymbol]) } def self.variants end end - class TruncationStrategy < OpenAI::BaseModel + class TruncationStrategy < OpenAI::Internal::Type::BaseModel # The truncation strategy to use for the thread. The default is `auto`. If set to # `last_messages`, the thread will be truncated to the n most recent messages in # the thread. When set to `auto`, messages in the middle of the thread will be @@ -623,7 +623,7 @@ module OpenAI # the thread. When set to `auto`, messages in the middle of the thread will be # dropped to fit the context length of the model, `max_prompt_tokens`. module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::RunCreateParams::TruncationStrategy::Type) } diff --git a/rbi/lib/openai/models/beta/threads/run_list_params.rbi b/rbi/lib/openai/models/beta/threads/run_list_params.rbi index da3c8bc7..17f72cc4 100644 --- a/rbi/lib/openai/models/beta/threads/run_list_params.rbi +++ b/rbi/lib/openai/models/beta/threads/run_list_params.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class RunListParams < OpenAI::BaseModel + class RunListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -50,7 +50,7 @@ module OpenAI before: String, limit: Integer, order: OpenAI::Models::Beta::Threads::RunListParams::Order::OrSymbol, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -75,7 +75,7 @@ module OpenAI # Sort order by the `created_at` timestamp of the objects. `asc` for ascending # order and `desc` for descending order. module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::RunListParams::Order) } OrSymbol = diff --git a/rbi/lib/openai/models/beta/threads/run_retrieve_params.rbi b/rbi/lib/openai/models/beta/threads/run_retrieve_params.rbi index f976ec26..60f2a769 100644 --- a/rbi/lib/openai/models/beta/threads/run_retrieve_params.rbi +++ b/rbi/lib/openai/models/beta/threads/run_retrieve_params.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class RunRetrieveParams < OpenAI::BaseModel + class RunRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -16,7 +16,7 @@ module OpenAI thread_id: String, request_options: T.any( OpenAI::RequestOptions, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .returns(T.attached_class) diff --git a/rbi/lib/openai/models/beta/threads/run_status.rbi b/rbi/lib/openai/models/beta/threads/run_status.rbi index 59df355d..e9d84d19 100644 --- a/rbi/lib/openai/models/beta/threads/run_status.rbi +++ b/rbi/lib/openai/models/beta/threads/run_status.rbi @@ -8,7 +8,7 @@ module OpenAI # `requires_action`, `cancelling`, `cancelled`, `failed`, `completed`, # `incomplete`, or `expired`. module RunStatus - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::RunStatus) } OrSymbol = T.type_alias { T.any(Symbol, String, OpenAI::Models::Beta::Threads::RunStatus::TaggedSymbol) } diff --git a/rbi/lib/openai/models/beta/threads/run_submit_tool_outputs_params.rbi b/rbi/lib/openai/models/beta/threads/run_submit_tool_outputs_params.rbi index 43e56a5e..1c8321a5 100644 --- a/rbi/lib/openai/models/beta/threads/run_submit_tool_outputs_params.rbi +++ b/rbi/lib/openai/models/beta/threads/run_submit_tool_outputs_params.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class RunSubmitToolOutputsParams < OpenAI::BaseModel + class RunSubmitToolOutputsParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -18,13 +18,8 @@ module OpenAI sig do params( thread_id: String, - tool_outputs: T::Array[ - T.any( - OpenAI::Models::Beta::Threads::RunSubmitToolOutputsParams::ToolOutput, - OpenAI::Internal::Util::AnyHash - ) - ], - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + tool_outputs: T::Array[T.any(OpenAI::Models::Beta::Threads::RunSubmitToolOutputsParams::ToolOutput, OpenAI::Internal::AnyHash)], + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -44,7 +39,7 @@ module OpenAI def to_hash end - class ToolOutput < OpenAI::BaseModel + class ToolOutput < OpenAI::Internal::Type::BaseModel # The output of the tool call to be submitted to continue the run. sig { returns(T.nilable(String)) } attr_reader :output diff --git a/rbi/lib/openai/models/beta/threads/run_update_params.rbi b/rbi/lib/openai/models/beta/threads/run_update_params.rbi index d4946a6e..a1a1a651 100644 --- a/rbi/lib/openai/models/beta/threads/run_update_params.rbi +++ b/rbi/lib/openai/models/beta/threads/run_update_params.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class RunUpdateParams < OpenAI::BaseModel + class RunUpdateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -24,7 +24,7 @@ module OpenAI params( thread_id: String, metadata: T.nilable(T::Hash[Symbol, String]), - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end diff --git a/rbi/lib/openai/models/beta/threads/runs/code_interpreter_logs.rbi b/rbi/lib/openai/models/beta/threads/runs/code_interpreter_logs.rbi index 76b660b5..20f4803a 100644 --- a/rbi/lib/openai/models/beta/threads/runs/code_interpreter_logs.rbi +++ b/rbi/lib/openai/models/beta/threads/runs/code_interpreter_logs.rbi @@ -5,7 +5,7 @@ module OpenAI module Beta module Threads module Runs - class CodeInterpreterLogs < OpenAI::BaseModel + class CodeInterpreterLogs < OpenAI::Internal::Type::BaseModel # The index of the output in the outputs array. sig { returns(Integer) } attr_accessor :index diff --git a/rbi/lib/openai/models/beta/threads/runs/code_interpreter_output_image.rbi b/rbi/lib/openai/models/beta/threads/runs/code_interpreter_output_image.rbi index 3b1be064..bff183df 100644 --- a/rbi/lib/openai/models/beta/threads/runs/code_interpreter_output_image.rbi +++ b/rbi/lib/openai/models/beta/threads/runs/code_interpreter_output_image.rbi @@ -5,7 +5,7 @@ module OpenAI module Beta module Threads module Runs - class CodeInterpreterOutputImage < OpenAI::BaseModel + class CodeInterpreterOutputImage < OpenAI::Internal::Type::BaseModel # The index of the output in the outputs array. sig { returns(Integer) } attr_accessor :index @@ -19,10 +19,7 @@ module OpenAI sig do params( - image: T.any( - OpenAI::Models::Beta::Threads::Runs::CodeInterpreterOutputImage::Image, - OpenAI::Internal::Util::AnyHash - ) + image: T.any(OpenAI::Models::Beta::Threads::Runs::CodeInterpreterOutputImage::Image, OpenAI::Internal::AnyHash) ) .void end @@ -31,10 +28,7 @@ module OpenAI sig do params( index: Integer, - image: T.any( - OpenAI::Models::Beta::Threads::Runs::CodeInterpreterOutputImage::Image, - OpenAI::Internal::Util::AnyHash - ), + image: T.any(OpenAI::Models::Beta::Threads::Runs::CodeInterpreterOutputImage::Image, OpenAI::Internal::AnyHash), type: Symbol ) .returns(T.attached_class) @@ -51,7 +45,7 @@ module OpenAI def to_hash end - class Image < OpenAI::BaseModel + class Image < OpenAI::Internal::Type::BaseModel # The [file](https://platform.openai.com/docs/api-reference/files) ID of the # image. sig { returns(T.nilable(String)) } diff --git a/rbi/lib/openai/models/beta/threads/runs/code_interpreter_tool_call.rbi b/rbi/lib/openai/models/beta/threads/runs/code_interpreter_tool_call.rbi index f0524366..dcf372d2 100644 --- a/rbi/lib/openai/models/beta/threads/runs/code_interpreter_tool_call.rbi +++ b/rbi/lib/openai/models/beta/threads/runs/code_interpreter_tool_call.rbi @@ -5,7 +5,7 @@ module OpenAI module Beta module Threads module Runs - class CodeInterpreterToolCall < OpenAI::BaseModel + class CodeInterpreterToolCall < OpenAI::Internal::Type::BaseModel # The ID of the tool call. sig { returns(String) } attr_accessor :id @@ -18,7 +18,7 @@ module OpenAI params( code_interpreter: T.any( OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCall::CodeInterpreter, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .void @@ -36,7 +36,7 @@ module OpenAI id: String, code_interpreter: T.any( OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCall::CodeInterpreter, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ), type: Symbol ) @@ -58,7 +58,7 @@ module OpenAI def to_hash end - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel # The input to the Code Interpreter tool call. sig { returns(String) } attr_accessor :input @@ -85,7 +85,7 @@ module OpenAI outputs: T::Array[ T.any( OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCall::CodeInterpreter::Output::Logs, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCall::CodeInterpreter::Output::Image ) ] @@ -114,9 +114,9 @@ module OpenAI # Text output from the Code Interpreter tool call as part of a run step. module Output - extend OpenAI::Union + extend OpenAI::Internal::Type::Union - class Logs < OpenAI::BaseModel + class Logs < OpenAI::Internal::Type::BaseModel # The text output from the Code Interpreter tool call. sig { returns(String) } attr_accessor :logs @@ -135,7 +135,7 @@ module OpenAI end end - class Image < OpenAI::BaseModel + class Image < OpenAI::Internal::Type::BaseModel sig do returns( OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCall::CodeInterpreter::Output::Image::Image @@ -147,7 +147,7 @@ module OpenAI params( image: T.any( OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCall::CodeInterpreter::Output::Image::Image, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .void @@ -162,7 +162,7 @@ module OpenAI params( image: T.any( OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCall::CodeInterpreter::Output::Image::Image, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ), type: Symbol ) @@ -183,7 +183,7 @@ module OpenAI def to_hash end - class Image < OpenAI::BaseModel + class Image < OpenAI::Internal::Type::BaseModel # The [file](https://platform.openai.com/docs/api-reference/files) ID of the # image. sig { returns(String) } diff --git a/rbi/lib/openai/models/beta/threads/runs/code_interpreter_tool_call_delta.rbi b/rbi/lib/openai/models/beta/threads/runs/code_interpreter_tool_call_delta.rbi index 999ac980..a3b42808 100644 --- a/rbi/lib/openai/models/beta/threads/runs/code_interpreter_tool_call_delta.rbi +++ b/rbi/lib/openai/models/beta/threads/runs/code_interpreter_tool_call_delta.rbi @@ -5,7 +5,7 @@ module OpenAI module Beta module Threads module Runs - class CodeInterpreterToolCallDelta < OpenAI::BaseModel + class CodeInterpreterToolCallDelta < OpenAI::Internal::Type::BaseModel # The index of the tool call in the tool calls array. sig { returns(Integer) } attr_accessor :index @@ -30,7 +30,7 @@ module OpenAI params( code_interpreter: T.any( OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCallDelta::CodeInterpreter, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .void @@ -44,7 +44,7 @@ module OpenAI id: String, code_interpreter: T.any( OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCallDelta::CodeInterpreter, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ), type: Symbol ) @@ -67,7 +67,7 @@ module OpenAI def to_hash end - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel # The input to the Code Interpreter tool call. sig { returns(T.nilable(String)) } attr_reader :input @@ -97,7 +97,7 @@ module OpenAI outputs: T::Array[ T.any( OpenAI::Models::Beta::Threads::Runs::CodeInterpreterLogs, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::Runs::CodeInterpreterOutputImage ) ] @@ -113,7 +113,7 @@ module OpenAI outputs: T::Array[ T.any( OpenAI::Models::Beta::Threads::Runs::CodeInterpreterLogs, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::Runs::CodeInterpreterOutputImage ) ] @@ -142,7 +142,7 @@ module OpenAI # Text output from the Code Interpreter tool call as part of a run step. module Output - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/beta/threads/runs/file_search_tool_call.rbi b/rbi/lib/openai/models/beta/threads/runs/file_search_tool_call.rbi index cbbf08ee..bdd404fb 100644 --- a/rbi/lib/openai/models/beta/threads/runs/file_search_tool_call.rbi +++ b/rbi/lib/openai/models/beta/threads/runs/file_search_tool_call.rbi @@ -5,7 +5,7 @@ module OpenAI module Beta module Threads module Runs - class FileSearchToolCall < OpenAI::BaseModel + class FileSearchToolCall < OpenAI::Internal::Type::BaseModel # The ID of the tool call object. sig { returns(String) } attr_accessor :id @@ -16,10 +16,7 @@ module OpenAI sig do params( - file_search: T.any( - OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch, - OpenAI::Internal::Util::AnyHash - ) + file_search: T.any(OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch, OpenAI::Internal::AnyHash) ) .void end @@ -33,10 +30,7 @@ module OpenAI sig do params( id: String, - file_search: T.any( - OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch, - OpenAI::Internal::Util::AnyHash - ), + file_search: T.any(OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch, OpenAI::Internal::AnyHash), type: Symbol ) .returns(T.attached_class) @@ -53,7 +47,7 @@ module OpenAI def to_hash end - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel # The ranking options for the file search. sig { returns(T.nilable(OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch::RankingOptions)) } attr_reader :ranking_options @@ -62,7 +56,7 @@ module OpenAI params( ranking_options: T.any( OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch::RankingOptions, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .void @@ -78,7 +72,7 @@ module OpenAI results: T::Array[ T.any( OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch::Result, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ] ) @@ -91,12 +85,12 @@ module OpenAI params( ranking_options: T.any( OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch::RankingOptions, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ), results: T::Array[ T.any( OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch::Result, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ] ) @@ -117,7 +111,7 @@ module OpenAI def to_hash end - class RankingOptions < OpenAI::BaseModel + class RankingOptions < OpenAI::Internal::Type::BaseModel # The ranker to use for the file search. If not specified will use the `auto` # ranker. sig do @@ -158,7 +152,7 @@ module OpenAI # The ranker to use for the file search. If not specified will use the `auto` # ranker. module Ranker - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch::RankingOptions::Ranker) } @@ -193,7 +187,7 @@ module OpenAI end end - class Result < OpenAI::BaseModel + class Result < OpenAI::Internal::Type::BaseModel # The ID of the file that result was found in. sig { returns(String) } attr_accessor :file_id @@ -221,7 +215,7 @@ module OpenAI content: T::Array[ T.any( OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch::Result::Content, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ] ) @@ -238,7 +232,7 @@ module OpenAI content: T::Array[ T.any( OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch::Result::Content, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ] ) @@ -261,7 +255,7 @@ module OpenAI def to_hash end - class Content < OpenAI::BaseModel + class Content < OpenAI::Internal::Type::BaseModel # The text content of the file. sig { returns(T.nilable(String)) } attr_reader :text @@ -311,7 +305,7 @@ module OpenAI # The type of the content. module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch::Result::Content::Type) } diff --git a/rbi/lib/openai/models/beta/threads/runs/file_search_tool_call_delta.rbi b/rbi/lib/openai/models/beta/threads/runs/file_search_tool_call_delta.rbi index 8dda0ce6..5474e165 100644 --- a/rbi/lib/openai/models/beta/threads/runs/file_search_tool_call_delta.rbi +++ b/rbi/lib/openai/models/beta/threads/runs/file_search_tool_call_delta.rbi @@ -5,7 +5,7 @@ module OpenAI module Beta module Threads module Runs - class FileSearchToolCallDelta < OpenAI::BaseModel + class FileSearchToolCallDelta < OpenAI::Internal::Type::BaseModel # For now, this is always going to be an empty object. sig { returns(T.anything) } attr_accessor :file_search diff --git a/rbi/lib/openai/models/beta/threads/runs/function_tool_call.rbi b/rbi/lib/openai/models/beta/threads/runs/function_tool_call.rbi index f4189ea3..2d1aa5a6 100644 --- a/rbi/lib/openai/models/beta/threads/runs/function_tool_call.rbi +++ b/rbi/lib/openai/models/beta/threads/runs/function_tool_call.rbi @@ -5,7 +5,7 @@ module OpenAI module Beta module Threads module Runs - class FunctionToolCall < OpenAI::BaseModel + class FunctionToolCall < OpenAI::Internal::Type::BaseModel # The ID of the tool call object. sig { returns(String) } attr_accessor :id @@ -16,7 +16,7 @@ module OpenAI sig do params( - function: T.any(OpenAI::Models::Beta::Threads::Runs::FunctionToolCall::Function, OpenAI::Internal::Util::AnyHash) + function: T.any(OpenAI::Models::Beta::Threads::Runs::FunctionToolCall::Function, OpenAI::Internal::AnyHash) ) .void end @@ -30,7 +30,7 @@ module OpenAI sig do params( id: String, - function: T.any(OpenAI::Models::Beta::Threads::Runs::FunctionToolCall::Function, OpenAI::Internal::Util::AnyHash), + function: T.any(OpenAI::Models::Beta::Threads::Runs::FunctionToolCall::Function, OpenAI::Internal::AnyHash), type: Symbol ) .returns(T.attached_class) @@ -47,7 +47,7 @@ module OpenAI def to_hash end - class Function < OpenAI::BaseModel + class Function < OpenAI::Internal::Type::BaseModel # The arguments passed to the function. sig { returns(String) } attr_accessor :arguments diff --git a/rbi/lib/openai/models/beta/threads/runs/function_tool_call_delta.rbi b/rbi/lib/openai/models/beta/threads/runs/function_tool_call_delta.rbi index b478742b..22b29632 100644 --- a/rbi/lib/openai/models/beta/threads/runs/function_tool_call_delta.rbi +++ b/rbi/lib/openai/models/beta/threads/runs/function_tool_call_delta.rbi @@ -5,7 +5,7 @@ module OpenAI module Beta module Threads module Runs - class FunctionToolCallDelta < OpenAI::BaseModel + class FunctionToolCallDelta < OpenAI::Internal::Type::BaseModel # The index of the tool call in the tool calls array. sig { returns(Integer) } attr_accessor :index @@ -28,10 +28,7 @@ module OpenAI sig do params( - function: T.any( - OpenAI::Models::Beta::Threads::Runs::FunctionToolCallDelta::Function, - OpenAI::Internal::Util::AnyHash - ) + function: T.any(OpenAI::Models::Beta::Threads::Runs::FunctionToolCallDelta::Function, OpenAI::Internal::AnyHash) ) .void end @@ -41,10 +38,7 @@ module OpenAI params( index: Integer, id: String, - function: T.any( - OpenAI::Models::Beta::Threads::Runs::FunctionToolCallDelta::Function, - OpenAI::Internal::Util::AnyHash - ), + function: T.any(OpenAI::Models::Beta::Threads::Runs::FunctionToolCallDelta::Function, OpenAI::Internal::AnyHash), type: Symbol ) .returns(T.attached_class) @@ -66,7 +60,7 @@ module OpenAI def to_hash end - class Function < OpenAI::BaseModel + class Function < OpenAI::Internal::Type::BaseModel # The arguments passed to the function. sig { returns(T.nilable(String)) } attr_reader :arguments diff --git a/rbi/lib/openai/models/beta/threads/runs/message_creation_step_details.rbi b/rbi/lib/openai/models/beta/threads/runs/message_creation_step_details.rbi index 2f3d4925..f279cb12 100644 --- a/rbi/lib/openai/models/beta/threads/runs/message_creation_step_details.rbi +++ b/rbi/lib/openai/models/beta/threads/runs/message_creation_step_details.rbi @@ -5,7 +5,7 @@ module OpenAI module Beta module Threads module Runs - class MessageCreationStepDetails < OpenAI::BaseModel + class MessageCreationStepDetails < OpenAI::Internal::Type::BaseModel sig { returns(OpenAI::Models::Beta::Threads::Runs::MessageCreationStepDetails::MessageCreation) } attr_reader :message_creation @@ -13,7 +13,7 @@ module OpenAI params( message_creation: T.any( OpenAI::Models::Beta::Threads::Runs::MessageCreationStepDetails::MessageCreation, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .void @@ -29,7 +29,7 @@ module OpenAI params( message_creation: T.any( OpenAI::Models::Beta::Threads::Runs::MessageCreationStepDetails::MessageCreation, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ), type: Symbol ) @@ -50,7 +50,7 @@ module OpenAI def to_hash end - class MessageCreation < OpenAI::BaseModel + class MessageCreation < OpenAI::Internal::Type::BaseModel # The ID of the message that was created by this run step. sig { returns(String) } attr_accessor :message_id diff --git a/rbi/lib/openai/models/beta/threads/runs/run_step.rbi b/rbi/lib/openai/models/beta/threads/runs/run_step.rbi index c068ad2f..af9b15a8 100644 --- a/rbi/lib/openai/models/beta/threads/runs/run_step.rbi +++ b/rbi/lib/openai/models/beta/threads/runs/run_step.rbi @@ -5,7 +5,7 @@ module OpenAI module Beta module Threads module Runs - class RunStep < OpenAI::BaseModel + class RunStep < OpenAI::Internal::Type::BaseModel # The identifier of the run step, which can be referenced in API endpoints. sig { returns(String) } attr_accessor :id @@ -44,7 +44,7 @@ module OpenAI sig do params( - last_error: T.nilable(T.any(OpenAI::Models::Beta::Threads::Runs::RunStep::LastError, OpenAI::Internal::Util::AnyHash)) + last_error: T.nilable(T.any(OpenAI::Models::Beta::Threads::Runs::RunStep::LastError, OpenAI::Internal::AnyHash)) ) .void end @@ -100,7 +100,7 @@ module OpenAI sig do params( - usage: T.nilable(T.any(OpenAI::Models::Beta::Threads::Runs::RunStep::Usage, OpenAI::Internal::Util::AnyHash)) + usage: T.nilable(T.any(OpenAI::Models::Beta::Threads::Runs::RunStep::Usage, OpenAI::Internal::AnyHash)) ) .void end @@ -116,18 +116,18 @@ module OpenAI created_at: Integer, expired_at: T.nilable(Integer), failed_at: T.nilable(Integer), - last_error: T.nilable(T.any(OpenAI::Models::Beta::Threads::Runs::RunStep::LastError, OpenAI::Internal::Util::AnyHash)), + last_error: T.nilable(T.any(OpenAI::Models::Beta::Threads::Runs::RunStep::LastError, OpenAI::Internal::AnyHash)), metadata: T.nilable(T::Hash[Symbol, String]), run_id: String, status: OpenAI::Models::Beta::Threads::Runs::RunStep::Status::OrSymbol, step_details: T.any( OpenAI::Models::Beta::Threads::Runs::MessageCreationStepDetails, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::Runs::ToolCallsStepDetails ), thread_id: String, type: OpenAI::Models::Beta::Threads::Runs::RunStep::Type::OrSymbol, - usage: T.nilable(T.any(OpenAI::Models::Beta::Threads::Runs::RunStep::Usage, OpenAI::Internal::Util::AnyHash)), + usage: T.nilable(T.any(OpenAI::Models::Beta::Threads::Runs::RunStep::Usage, OpenAI::Internal::AnyHash)), object: Symbol ) .returns(T.attached_class) @@ -181,7 +181,7 @@ module OpenAI def to_hash end - class LastError < OpenAI::BaseModel + class LastError < OpenAI::Internal::Type::BaseModel # One of `server_error` or `rate_limit_exceeded`. sig { returns(OpenAI::Models::Beta::Threads::Runs::RunStep::LastError::Code::TaggedSymbol) } attr_accessor :code @@ -210,7 +210,7 @@ module OpenAI # One of `server_error` or `rate_limit_exceeded`. module Code - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::Runs::RunStep::LastError::Code) } @@ -231,7 +231,7 @@ module OpenAI # The status of the run step, which can be either `in_progress`, `cancelled`, # `failed`, `completed`, or `expired`. module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::Runs::RunStep::Status) } OrSymbol = @@ -250,7 +250,7 @@ module OpenAI # The details of the run step. module StepDetails - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override @@ -264,7 +264,7 @@ module OpenAI # The type of run step, which can be either `message_creation` or `tool_calls`. module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::Runs::RunStep::Type) } OrSymbol = @@ -279,7 +279,7 @@ module OpenAI end end - class Usage < OpenAI::BaseModel + class Usage < OpenAI::Internal::Type::BaseModel # Number of completion tokens used over the course of the run step. sig { returns(Integer) } attr_accessor :completion_tokens diff --git a/rbi/lib/openai/models/beta/threads/runs/run_step_delta.rbi b/rbi/lib/openai/models/beta/threads/runs/run_step_delta.rbi index fa423211..a599ca89 100644 --- a/rbi/lib/openai/models/beta/threads/runs/run_step_delta.rbi +++ b/rbi/lib/openai/models/beta/threads/runs/run_step_delta.rbi @@ -5,7 +5,7 @@ module OpenAI module Beta module Threads module Runs - class RunStepDelta < OpenAI::BaseModel + class RunStepDelta < OpenAI::Internal::Type::BaseModel # The details of the run step. sig do returns( @@ -23,7 +23,7 @@ module OpenAI params( step_details: T.any( OpenAI::Models::Beta::Threads::Runs::RunStepDeltaMessageDelta, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::Runs::ToolCallDeltaObject ) ) @@ -36,7 +36,7 @@ module OpenAI params( step_details: T.any( OpenAI::Models::Beta::Threads::Runs::RunStepDeltaMessageDelta, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::Runs::ToolCallDeltaObject ) ) @@ -61,7 +61,7 @@ module OpenAI # The details of the run step. module StepDetails - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/beta/threads/runs/run_step_delta_event.rbi b/rbi/lib/openai/models/beta/threads/runs/run_step_delta_event.rbi index 1369d183..fbaa33df 100644 --- a/rbi/lib/openai/models/beta/threads/runs/run_step_delta_event.rbi +++ b/rbi/lib/openai/models/beta/threads/runs/run_step_delta_event.rbi @@ -5,7 +5,7 @@ module OpenAI module Beta module Threads module Runs - class RunStepDeltaEvent < OpenAI::BaseModel + class RunStepDeltaEvent < OpenAI::Internal::Type::BaseModel # The identifier of the run step, which can be referenced in API endpoints. sig { returns(String) } attr_accessor :id @@ -14,10 +14,7 @@ module OpenAI sig { returns(OpenAI::Models::Beta::Threads::Runs::RunStepDelta) } attr_reader :delta - sig do - params(delta: T.any(OpenAI::Models::Beta::Threads::Runs::RunStepDelta, OpenAI::Internal::Util::AnyHash)) - .void - end + sig { params(delta: T.any(OpenAI::Models::Beta::Threads::Runs::RunStepDelta, OpenAI::Internal::AnyHash)).void } attr_writer :delta # The object type, which is always `thread.run.step.delta`. @@ -29,7 +26,7 @@ module OpenAI sig do params( id: String, - delta: T.any(OpenAI::Models::Beta::Threads::Runs::RunStepDelta, OpenAI::Internal::Util::AnyHash), + delta: T.any(OpenAI::Models::Beta::Threads::Runs::RunStepDelta, OpenAI::Internal::AnyHash), object: Symbol ) .returns(T.attached_class) diff --git a/rbi/lib/openai/models/beta/threads/runs/run_step_delta_message_delta.rbi b/rbi/lib/openai/models/beta/threads/runs/run_step_delta_message_delta.rbi index 3408ca06..d4dd14e5 100644 --- a/rbi/lib/openai/models/beta/threads/runs/run_step_delta_message_delta.rbi +++ b/rbi/lib/openai/models/beta/threads/runs/run_step_delta_message_delta.rbi @@ -5,7 +5,7 @@ module OpenAI module Beta module Threads module Runs - class RunStepDeltaMessageDelta < OpenAI::BaseModel + class RunStepDeltaMessageDelta < OpenAI::Internal::Type::BaseModel # Always `message_creation`. sig { returns(Symbol) } attr_accessor :type @@ -17,7 +17,7 @@ module OpenAI params( message_creation: T.any( OpenAI::Models::Beta::Threads::Runs::RunStepDeltaMessageDelta::MessageCreation, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .void @@ -29,7 +29,7 @@ module OpenAI params( message_creation: T.any( OpenAI::Models::Beta::Threads::Runs::RunStepDeltaMessageDelta::MessageCreation, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ), type: Symbol ) @@ -50,7 +50,7 @@ module OpenAI def to_hash end - class MessageCreation < OpenAI::BaseModel + class MessageCreation < OpenAI::Internal::Type::BaseModel # The ID of the message that was created by this run step. sig { returns(T.nilable(String)) } attr_reader :message_id diff --git a/rbi/lib/openai/models/beta/threads/runs/run_step_include.rbi b/rbi/lib/openai/models/beta/threads/runs/run_step_include.rbi index 8741e9f7..bee1919e 100644 --- a/rbi/lib/openai/models/beta/threads/runs/run_step_include.rbi +++ b/rbi/lib/openai/models/beta/threads/runs/run_step_include.rbi @@ -6,7 +6,7 @@ module OpenAI module Threads module Runs module RunStepInclude - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::Runs::RunStepInclude) } OrSymbol = diff --git a/rbi/lib/openai/models/beta/threads/runs/step_list_params.rbi b/rbi/lib/openai/models/beta/threads/runs/step_list_params.rbi index 6b012ab4..e48d41bb 100644 --- a/rbi/lib/openai/models/beta/threads/runs/step_list_params.rbi +++ b/rbi/lib/openai/models/beta/threads/runs/step_list_params.rbi @@ -5,7 +5,7 @@ module OpenAI module Beta module Threads module Runs - class StepListParams < OpenAI::BaseModel + class StepListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -69,7 +69,7 @@ module OpenAI include: T::Array[OpenAI::Models::Beta::Threads::Runs::RunStepInclude::OrSymbol], limit: Integer, order: OpenAI::Models::Beta::Threads::Runs::StepListParams::Order::OrSymbol, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -104,7 +104,7 @@ module OpenAI # Sort order by the `created_at` timestamp of the objects. `asc` for ascending # order and `desc` for descending order. module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Beta::Threads::Runs::StepListParams::Order) } OrSymbol = diff --git a/rbi/lib/openai/models/beta/threads/runs/step_retrieve_params.rbi b/rbi/lib/openai/models/beta/threads/runs/step_retrieve_params.rbi index fd32d2e8..bd43d059 100644 --- a/rbi/lib/openai/models/beta/threads/runs/step_retrieve_params.rbi +++ b/rbi/lib/openai/models/beta/threads/runs/step_retrieve_params.rbi @@ -5,7 +5,7 @@ module OpenAI module Beta module Threads module Runs - class StepRetrieveParams < OpenAI::BaseModel + class StepRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -33,7 +33,7 @@ module OpenAI thread_id: String, run_id: String, include: T::Array[OpenAI::Models::Beta::Threads::Runs::RunStepInclude::OrSymbol], - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end diff --git a/rbi/lib/openai/models/beta/threads/runs/tool_call.rbi b/rbi/lib/openai/models/beta/threads/runs/tool_call.rbi index 22b596d5..8bdd948f 100644 --- a/rbi/lib/openai/models/beta/threads/runs/tool_call.rbi +++ b/rbi/lib/openai/models/beta/threads/runs/tool_call.rbi @@ -7,7 +7,7 @@ module OpenAI module Runs # Details of the Code Interpreter tool call the run step was involved in. module ToolCall - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/beta/threads/runs/tool_call_delta.rbi b/rbi/lib/openai/models/beta/threads/runs/tool_call_delta.rbi index 8af6cb90..a1e18b64 100644 --- a/rbi/lib/openai/models/beta/threads/runs/tool_call_delta.rbi +++ b/rbi/lib/openai/models/beta/threads/runs/tool_call_delta.rbi @@ -7,7 +7,7 @@ module OpenAI module Runs # Details of the Code Interpreter tool call the run step was involved in. module ToolCallDelta - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/beta/threads/runs/tool_call_delta_object.rbi b/rbi/lib/openai/models/beta/threads/runs/tool_call_delta_object.rbi index 783946aa..d9051b8e 100644 --- a/rbi/lib/openai/models/beta/threads/runs/tool_call_delta_object.rbi +++ b/rbi/lib/openai/models/beta/threads/runs/tool_call_delta_object.rbi @@ -5,7 +5,7 @@ module OpenAI module Beta module Threads module Runs - class ToolCallDeltaObject < OpenAI::BaseModel + class ToolCallDeltaObject < OpenAI::Internal::Type::BaseModel # Always `tool_calls`. sig { returns(Symbol) } attr_accessor :type @@ -33,7 +33,7 @@ module OpenAI tool_calls: T::Array[ T.any( OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCallDelta, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::Runs::FileSearchToolCallDelta, OpenAI::Models::Beta::Threads::Runs::FunctionToolCallDelta ) @@ -49,7 +49,7 @@ module OpenAI tool_calls: T::Array[ T.any( OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCallDelta, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::Runs::FileSearchToolCallDelta, OpenAI::Models::Beta::Threads::Runs::FunctionToolCallDelta ) diff --git a/rbi/lib/openai/models/beta/threads/runs/tool_calls_step_details.rbi b/rbi/lib/openai/models/beta/threads/runs/tool_calls_step_details.rbi index 3b5ca709..b85467c4 100644 --- a/rbi/lib/openai/models/beta/threads/runs/tool_calls_step_details.rbi +++ b/rbi/lib/openai/models/beta/threads/runs/tool_calls_step_details.rbi @@ -5,7 +5,7 @@ module OpenAI module Beta module Threads module Runs - class ToolCallsStepDetails < OpenAI::BaseModel + class ToolCallsStepDetails < OpenAI::Internal::Type::BaseModel # An array of tool calls the run step was involved in. These can be associated # with one of three types of tools: `code_interpreter`, `file_search`, or # `function`. @@ -32,7 +32,7 @@ module OpenAI tool_calls: T::Array[ T.any( OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCall, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall, OpenAI::Models::Beta::Threads::Runs::FunctionToolCall ) diff --git a/rbi/lib/openai/models/beta/threads/text.rbi b/rbi/lib/openai/models/beta/threads/text.rbi index 96e9162b..933e4630 100644 --- a/rbi/lib/openai/models/beta/threads/text.rbi +++ b/rbi/lib/openai/models/beta/threads/text.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class Text < OpenAI::BaseModel + class Text < OpenAI::Internal::Type::BaseModel sig do returns( T::Array[ @@ -26,7 +26,7 @@ module OpenAI annotations: T::Array[ T.any( OpenAI::Models::Beta::Threads::FileCitationAnnotation, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::FilePathAnnotation ) ], diff --git a/rbi/lib/openai/models/beta/threads/text_content_block.rbi b/rbi/lib/openai/models/beta/threads/text_content_block.rbi index 2293760e..3838ff04 100644 --- a/rbi/lib/openai/models/beta/threads/text_content_block.rbi +++ b/rbi/lib/openai/models/beta/threads/text_content_block.rbi @@ -4,11 +4,11 @@ module OpenAI module Models module Beta module Threads - class TextContentBlock < OpenAI::BaseModel + class TextContentBlock < OpenAI::Internal::Type::BaseModel sig { returns(OpenAI::Models::Beta::Threads::Text) } attr_reader :text - sig { params(text: T.any(OpenAI::Models::Beta::Threads::Text, OpenAI::Internal::Util::AnyHash)).void } + sig { params(text: T.any(OpenAI::Models::Beta::Threads::Text, OpenAI::Internal::AnyHash)).void } attr_writer :text # Always `text`. @@ -17,7 +17,7 @@ module OpenAI # The text content that is part of a message. sig do - params(text: T.any(OpenAI::Models::Beta::Threads::Text, OpenAI::Internal::Util::AnyHash), type: Symbol) + params(text: T.any(OpenAI::Models::Beta::Threads::Text, OpenAI::Internal::AnyHash), type: Symbol) .returns(T.attached_class) end def self.new(text:, type: :text) diff --git a/rbi/lib/openai/models/beta/threads/text_content_block_param.rbi b/rbi/lib/openai/models/beta/threads/text_content_block_param.rbi index f59cbbd5..94f3431d 100644 --- a/rbi/lib/openai/models/beta/threads/text_content_block_param.rbi +++ b/rbi/lib/openai/models/beta/threads/text_content_block_param.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class TextContentBlockParam < OpenAI::BaseModel + class TextContentBlockParam < OpenAI::Internal::Type::BaseModel # Text content to be sent to the model sig { returns(String) } attr_accessor :text diff --git a/rbi/lib/openai/models/beta/threads/text_delta.rbi b/rbi/lib/openai/models/beta/threads/text_delta.rbi index a5aea637..8fda3e45 100644 --- a/rbi/lib/openai/models/beta/threads/text_delta.rbi +++ b/rbi/lib/openai/models/beta/threads/text_delta.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class TextDelta < OpenAI::BaseModel + class TextDelta < OpenAI::Internal::Type::BaseModel sig do returns( T.nilable( @@ -24,7 +24,7 @@ module OpenAI annotations: T::Array[ T.any( OpenAI::Models::Beta::Threads::FileCitationDeltaAnnotation, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::FilePathDeltaAnnotation ) ] @@ -45,7 +45,7 @@ module OpenAI annotations: T::Array[ T.any( OpenAI::Models::Beta::Threads::FileCitationDeltaAnnotation, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::FilePathDeltaAnnotation ) ], diff --git a/rbi/lib/openai/models/beta/threads/text_delta_block.rbi b/rbi/lib/openai/models/beta/threads/text_delta_block.rbi index 8d353d81..d3d54de5 100644 --- a/rbi/lib/openai/models/beta/threads/text_delta_block.rbi +++ b/rbi/lib/openai/models/beta/threads/text_delta_block.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Beta module Threads - class TextDeltaBlock < OpenAI::BaseModel + class TextDeltaBlock < OpenAI::Internal::Type::BaseModel # The index of the content part in the message. sig { returns(Integer) } attr_accessor :index @@ -16,14 +16,14 @@ module OpenAI sig { returns(T.nilable(OpenAI::Models::Beta::Threads::TextDelta)) } attr_reader :text - sig { params(text: T.any(OpenAI::Models::Beta::Threads::TextDelta, OpenAI::Internal::Util::AnyHash)).void } + sig { params(text: T.any(OpenAI::Models::Beta::Threads::TextDelta, OpenAI::Internal::AnyHash)).void } attr_writer :text # The text content that is part of a message. sig do params( index: Integer, - text: T.any(OpenAI::Models::Beta::Threads::TextDelta, OpenAI::Internal::Util::AnyHash), + text: T.any(OpenAI::Models::Beta::Threads::TextDelta, OpenAI::Internal::AnyHash), type: Symbol ) .returns(T.attached_class) diff --git a/rbi/lib/openai/models/chat/chat_completion.rbi b/rbi/lib/openai/models/chat/chat_completion.rbi index 7086df0f..829cff8c 100644 --- a/rbi/lib/openai/models/chat/chat_completion.rbi +++ b/rbi/lib/openai/models/chat/chat_completion.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletion < OpenAI::BaseModel + class ChatCompletion < OpenAI::Internal::Type::BaseModel # A unique identifier for the chat completion. sig { returns(String) } attr_accessor :id @@ -43,7 +43,7 @@ module OpenAI sig { returns(T.nilable(OpenAI::Models::CompletionUsage)) } attr_reader :usage - sig { params(usage: T.any(OpenAI::Models::CompletionUsage, OpenAI::Internal::Util::AnyHash)).void } + sig { params(usage: T.any(OpenAI::Models::CompletionUsage, OpenAI::Internal::AnyHash)).void } attr_writer :usage # Represents a chat completion response returned by model, based on the provided @@ -51,12 +51,12 @@ module OpenAI sig do params( id: String, - choices: T::Array[T.any(OpenAI::Models::Chat::ChatCompletion::Choice, OpenAI::Internal::Util::AnyHash)], + choices: T::Array[T.any(OpenAI::Models::Chat::ChatCompletion::Choice, OpenAI::Internal::AnyHash)], created: Integer, model: String, service_tier: T.nilable(OpenAI::Models::Chat::ChatCompletion::ServiceTier::OrSymbol), system_fingerprint: String, - usage: T.any(OpenAI::Models::CompletionUsage, OpenAI::Internal::Util::AnyHash), + usage: T.any(OpenAI::Models::CompletionUsage, OpenAI::Internal::AnyHash), object: Symbol ) .returns(T.attached_class) @@ -91,7 +91,7 @@ module OpenAI def to_hash end - class Choice < OpenAI::BaseModel + class Choice < OpenAI::Internal::Type::BaseModel # The reason the model stopped generating tokens. This will be `stop` if the model # hit a natural stop point or a provided stop sequence, `length` if the maximum # number of tokens specified in the request was reached, `content_filter` if @@ -111,7 +111,7 @@ module OpenAI sig do params( - logprobs: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletion::Choice::Logprobs, OpenAI::Internal::Util::AnyHash)) + logprobs: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletion::Choice::Logprobs, OpenAI::Internal::AnyHash)) ) .void end @@ -121,15 +121,15 @@ module OpenAI sig { returns(OpenAI::Models::Chat::ChatCompletionMessage) } attr_reader :message - sig { params(message: T.any(OpenAI::Models::Chat::ChatCompletionMessage, OpenAI::Internal::Util::AnyHash)).void } + sig { params(message: T.any(OpenAI::Models::Chat::ChatCompletionMessage, OpenAI::Internal::AnyHash)).void } attr_writer :message sig do params( finish_reason: OpenAI::Models::Chat::ChatCompletion::Choice::FinishReason::OrSymbol, index: Integer, - logprobs: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletion::Choice::Logprobs, OpenAI::Internal::Util::AnyHash)), - message: T.any(OpenAI::Models::Chat::ChatCompletionMessage, OpenAI::Internal::Util::AnyHash) + logprobs: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletion::Choice::Logprobs, OpenAI::Internal::AnyHash)), + message: T.any(OpenAI::Models::Chat::ChatCompletionMessage, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -157,7 +157,7 @@ module OpenAI # model called a tool, or `function_call` (deprecated) if the model called a # function. module FinishReason - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Chat::ChatCompletion::Choice::FinishReason) } OrSymbol = @@ -176,7 +176,7 @@ module OpenAI end end - class Logprobs < OpenAI::BaseModel + class Logprobs < OpenAI::Internal::Type::BaseModel # A list of message content tokens with log probability information. sig { returns(T.nilable(T::Array[OpenAI::Models::Chat::ChatCompletionTokenLogprob])) } attr_accessor :content @@ -188,12 +188,8 @@ module OpenAI # Log probability information for the choice. sig do params( - content: T.nilable( - T::Array[T.any(OpenAI::Models::Chat::ChatCompletionTokenLogprob, OpenAI::Internal::Util::AnyHash)] - ), - refusal: T.nilable( - T::Array[T.any(OpenAI::Models::Chat::ChatCompletionTokenLogprob, OpenAI::Internal::Util::AnyHash)] - ) + content: T.nilable(T::Array[T.any(OpenAI::Models::Chat::ChatCompletionTokenLogprob, OpenAI::Internal::AnyHash)]), + refusal: T.nilable(T::Array[T.any(OpenAI::Models::Chat::ChatCompletionTokenLogprob, OpenAI::Internal::AnyHash)]) ) .returns(T.attached_class) end @@ -216,7 +212,7 @@ module OpenAI # The service tier used for processing the request. module ServiceTier - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Chat::ChatCompletion::ServiceTier) } OrSymbol = diff --git a/rbi/lib/openai/models/chat/chat_completion_assistant_message_param.rbi b/rbi/lib/openai/models/chat/chat_completion_assistant_message_param.rbi index 40188371..a6f8cec6 100644 --- a/rbi/lib/openai/models/chat/chat_completion_assistant_message_param.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_assistant_message_param.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionAssistantMessageParam < OpenAI::BaseModel + class ChatCompletionAssistantMessageParam < OpenAI::Internal::Type::BaseModel # The role of the messages author, in this case `assistant`. sig { returns(Symbol) } attr_accessor :role @@ -16,7 +16,7 @@ module OpenAI sig do params( audio: T.nilable( - T.any(OpenAI::Models::Chat::ChatCompletionAssistantMessageParam::Audio, OpenAI::Internal::Util::AnyHash) + T.any(OpenAI::Models::Chat::ChatCompletionAssistantMessageParam::Audio, OpenAI::Internal::AnyHash) ) ) .void @@ -50,10 +50,7 @@ module OpenAI sig do params( function_call: T.nilable( - T.any( - OpenAI::Models::Chat::ChatCompletionAssistantMessageParam::FunctionCall, - OpenAI::Internal::Util::AnyHash - ) + T.any(OpenAI::Models::Chat::ChatCompletionAssistantMessageParam::FunctionCall, OpenAI::Internal::AnyHash) ) ) .void @@ -78,7 +75,7 @@ module OpenAI sig do params( - tool_calls: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionMessageToolCall, OpenAI::Internal::Util::AnyHash)] + tool_calls: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionMessageToolCall, OpenAI::Internal::AnyHash)] ) .void end @@ -88,7 +85,7 @@ module OpenAI sig do params( audio: T.nilable( - T.any(OpenAI::Models::Chat::ChatCompletionAssistantMessageParam::Audio, OpenAI::Internal::Util::AnyHash) + T.any(OpenAI::Models::Chat::ChatCompletionAssistantMessageParam::Audio, OpenAI::Internal::AnyHash) ), content: T.nilable( T.any( @@ -96,21 +93,18 @@ module OpenAI T::Array[ T.any( OpenAI::Models::Chat::ChatCompletionContentPartText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Chat::ChatCompletionContentPartRefusal ) ] ) ), function_call: T.nilable( - T.any( - OpenAI::Models::Chat::ChatCompletionAssistantMessageParam::FunctionCall, - OpenAI::Internal::Util::AnyHash - ) + T.any(OpenAI::Models::Chat::ChatCompletionAssistantMessageParam::FunctionCall, OpenAI::Internal::AnyHash) ), name: String, refusal: T.nilable(String), - tool_calls: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionMessageToolCall, OpenAI::Internal::Util::AnyHash)], + tool_calls: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionMessageToolCall, OpenAI::Internal::AnyHash)], role: Symbol ) .returns(T.attached_class) @@ -153,7 +147,7 @@ module OpenAI def to_hash end - class Audio < OpenAI::BaseModel + class Audio < OpenAI::Internal::Type::BaseModel # Unique identifier for a previous audio response from the model. sig { returns(String) } attr_accessor :id @@ -172,12 +166,12 @@ module OpenAI # The contents of the assistant message. Required unless `tool_calls` or # `function_call` is specified. module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # Learn about # [text inputs](https://platform.openai.com/docs/guides/text-generation). module ArrayOfContentPart - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override @@ -208,12 +202,12 @@ module OpenAI ArrayOfContentPartArray = T.let( - OpenAI::ArrayOf[union: OpenAI::Models::Chat::ChatCompletionAssistantMessageParam::Content::ArrayOfContentPart], + OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Chat::ChatCompletionAssistantMessageParam::Content::ArrayOfContentPart], OpenAI::Internal::Type::Converter ) end - class FunctionCall < OpenAI::BaseModel + class FunctionCall < OpenAI::Internal::Type::BaseModel # The arguments to call the function with, as generated by the model in JSON # format. Note that the model does not always generate valid JSON, and may # hallucinate parameters not defined by your function schema. Validate the diff --git a/rbi/lib/openai/models/chat/chat_completion_audio.rbi b/rbi/lib/openai/models/chat/chat_completion_audio.rbi index 7cd9cc72..fa10bf84 100644 --- a/rbi/lib/openai/models/chat/chat_completion_audio.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_audio.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionAudio < OpenAI::BaseModel + class ChatCompletionAudio < OpenAI::Internal::Type::BaseModel # Unique identifier for this audio response. sig { returns(String) } attr_accessor :id diff --git a/rbi/lib/openai/models/chat/chat_completion_audio_param.rbi b/rbi/lib/openai/models/chat/chat_completion_audio_param.rbi index 938f0366..29b450f1 100644 --- a/rbi/lib/openai/models/chat/chat_completion_audio_param.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_audio_param.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionAudioParam < OpenAI::BaseModel + class ChatCompletionAudioParam < OpenAI::Internal::Type::BaseModel # Specifies the output audio format. Must be one of `wav`, `mp3`, `flac`, `opus`, # or `pcm16`. sig { returns(OpenAI::Models::Chat::ChatCompletionAudioParam::Format::OrSymbol) } @@ -42,7 +42,7 @@ module OpenAI # Specifies the output audio format. Must be one of `wav`, `mp3`, `flac`, `opus`, # or `pcm16`. module Format - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Chat::ChatCompletionAudioParam::Format) } OrSymbol = @@ -62,7 +62,7 @@ module OpenAI # The voice the model uses to respond. Supported voices are `alloy`, `ash`, # `ballad`, `coral`, `echo`, `sage`, and `shimmer`. module Voice - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, OpenAI::Models::Chat::ChatCompletionAudioParam::Voice::OrSymbol]) } def self.variants diff --git a/rbi/lib/openai/models/chat/chat_completion_chunk.rbi b/rbi/lib/openai/models/chat/chat_completion_chunk.rbi index 719f97f8..d93d792a 100644 --- a/rbi/lib/openai/models/chat/chat_completion_chunk.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_chunk.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionChunk < OpenAI::BaseModel + class ChatCompletionChunk < OpenAI::Internal::Type::BaseModel # A unique identifier for the chat completion. Each chunk has the same ID. sig { returns(String) } attr_accessor :id @@ -50,7 +50,7 @@ module OpenAI sig { returns(T.nilable(OpenAI::Models::CompletionUsage)) } attr_reader :usage - sig { params(usage: T.nilable(T.any(OpenAI::Models::CompletionUsage, OpenAI::Internal::Util::AnyHash))).void } + sig { params(usage: T.nilable(T.any(OpenAI::Models::CompletionUsage, OpenAI::Internal::AnyHash))).void } attr_writer :usage # Represents a streamed chunk of a chat completion response returned by the model, @@ -59,12 +59,12 @@ module OpenAI sig do params( id: String, - choices: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionChunk::Choice, OpenAI::Internal::Util::AnyHash)], + choices: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionChunk::Choice, OpenAI::Internal::AnyHash)], created: Integer, model: String, service_tier: T.nilable(OpenAI::Models::Chat::ChatCompletionChunk::ServiceTier::OrSymbol), system_fingerprint: String, - usage: T.nilable(T.any(OpenAI::Models::CompletionUsage, OpenAI::Internal::Util::AnyHash)), + usage: T.nilable(T.any(OpenAI::Models::CompletionUsage, OpenAI::Internal::AnyHash)), object: Symbol ) .returns(T.attached_class) @@ -99,15 +99,13 @@ module OpenAI def to_hash end - class Choice < OpenAI::BaseModel + class Choice < OpenAI::Internal::Type::BaseModel # A chat completion delta generated by streamed model responses. sig { returns(OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta) } attr_reader :delta sig do - params( - delta: T.any(OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta, OpenAI::Internal::Util::AnyHash) - ) + params(delta: T.any(OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta, OpenAI::Internal::AnyHash)) .void end attr_writer :delta @@ -131,9 +129,7 @@ module OpenAI sig do params( - logprobs: T.nilable( - T.any(OpenAI::Models::Chat::ChatCompletionChunk::Choice::Logprobs, OpenAI::Internal::Util::AnyHash) - ) + logprobs: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionChunk::Choice::Logprobs, OpenAI::Internal::AnyHash)) ) .void end @@ -141,12 +137,10 @@ module OpenAI sig do params( - delta: T.any(OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta, OpenAI::Internal::Util::AnyHash), + delta: T.any(OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta, OpenAI::Internal::AnyHash), finish_reason: T.nilable(OpenAI::Models::Chat::ChatCompletionChunk::Choice::FinishReason::OrSymbol), index: Integer, - logprobs: T.nilable( - T.any(OpenAI::Models::Chat::ChatCompletionChunk::Choice::Logprobs, OpenAI::Internal::Util::AnyHash) - ) + logprobs: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionChunk::Choice::Logprobs, OpenAI::Internal::AnyHash)) ) .returns(T.attached_class) end @@ -167,7 +161,7 @@ module OpenAI def to_hash end - class Delta < OpenAI::BaseModel + class Delta < OpenAI::Internal::Type::BaseModel # The contents of the chunk message. sig { returns(T.nilable(String)) } attr_accessor :content @@ -179,10 +173,7 @@ module OpenAI sig do params( - function_call: T.any( - OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta::FunctionCall, - OpenAI::Internal::Util::AnyHash - ) + function_call: T.any(OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta::FunctionCall, OpenAI::Internal::AnyHash) ) .void end @@ -204,7 +195,7 @@ module OpenAI sig do params( - tool_calls: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta::ToolCall, OpenAI::Internal::Util::AnyHash)] + tool_calls: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta::ToolCall, OpenAI::Internal::AnyHash)] ) .void end @@ -214,13 +205,10 @@ module OpenAI sig do params( content: T.nilable(String), - function_call: T.any( - OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta::FunctionCall, - OpenAI::Internal::Util::AnyHash - ), + function_call: T.any(OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta::FunctionCall, OpenAI::Internal::AnyHash), refusal: T.nilable(String), role: OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta::Role::OrSymbol, - tool_calls: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta::ToolCall, OpenAI::Internal::Util::AnyHash)] + tool_calls: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta::ToolCall, OpenAI::Internal::AnyHash)] ) .returns(T.attached_class) end @@ -242,7 +230,7 @@ module OpenAI def to_hash end - class FunctionCall < OpenAI::BaseModel + class FunctionCall < OpenAI::Internal::Type::BaseModel # The arguments to call the function with, as generated by the model in JSON # format. Note that the model does not always generate valid JSON, and may # hallucinate parameters not defined by your function schema. Validate the @@ -273,7 +261,7 @@ module OpenAI # The role of the author of this message. module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta::Role) } @@ -293,7 +281,7 @@ module OpenAI end end - class ToolCall < OpenAI::BaseModel + class ToolCall < OpenAI::Internal::Type::BaseModel sig { returns(Integer) } attr_accessor :index @@ -311,7 +299,7 @@ module OpenAI params( function: T.any( OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta::ToolCall::Function, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .void @@ -331,7 +319,7 @@ module OpenAI id: String, function: T.any( OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta::ToolCall::Function, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ), type: OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta::ToolCall::Type::OrSymbol ) @@ -354,7 +342,7 @@ module OpenAI def to_hash end - class Function < OpenAI::BaseModel + class Function < OpenAI::Internal::Type::BaseModel # The arguments to call the function with, as generated by the model in JSON # format. Note that the model does not always generate valid JSON, and may # hallucinate parameters not defined by your function schema. Validate the @@ -383,7 +371,7 @@ module OpenAI # The type of the tool. Currently, only `function` is supported. module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta::ToolCall::Type) } @@ -416,7 +404,7 @@ module OpenAI # model called a tool, or `function_call` (deprecated) if the model called a # function. module FinishReason - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Chat::ChatCompletionChunk::Choice::FinishReason) } @@ -437,7 +425,7 @@ module OpenAI end end - class Logprobs < OpenAI::BaseModel + class Logprobs < OpenAI::Internal::Type::BaseModel # A list of message content tokens with log probability information. sig { returns(T.nilable(T::Array[OpenAI::Models::Chat::ChatCompletionTokenLogprob])) } attr_accessor :content @@ -449,12 +437,8 @@ module OpenAI # Log probability information for the choice. sig do params( - content: T.nilable( - T::Array[T.any(OpenAI::Models::Chat::ChatCompletionTokenLogprob, OpenAI::Internal::Util::AnyHash)] - ), - refusal: T.nilable( - T::Array[T.any(OpenAI::Models::Chat::ChatCompletionTokenLogprob, OpenAI::Internal::Util::AnyHash)] - ) + content: T.nilable(T::Array[T.any(OpenAI::Models::Chat::ChatCompletionTokenLogprob, OpenAI::Internal::AnyHash)]), + refusal: T.nilable(T::Array[T.any(OpenAI::Models::Chat::ChatCompletionTokenLogprob, OpenAI::Internal::AnyHash)]) ) .returns(T.attached_class) end @@ -477,7 +461,7 @@ module OpenAI # The service tier used for processing the request. module ServiceTier - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Chat::ChatCompletionChunk::ServiceTier) } OrSymbol = diff --git a/rbi/lib/openai/models/chat/chat_completion_content_part.rbi b/rbi/lib/openai/models/chat/chat_completion_content_part.rbi index df2d9e47..f53e1e23 100644 --- a/rbi/lib/openai/models/chat/chat_completion_content_part.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_content_part.rbi @@ -6,15 +6,15 @@ module OpenAI # Learn about # [text inputs](https://platform.openai.com/docs/guides/text-generation). module ChatCompletionContentPart - extend OpenAI::Union + extend OpenAI::Internal::Type::Union - class File < OpenAI::BaseModel + class File < OpenAI::Internal::Type::BaseModel sig { returns(OpenAI::Models::Chat::ChatCompletionContentPart::File::File) } attr_reader :file sig do params( - file: T.any(OpenAI::Models::Chat::ChatCompletionContentPart::File::File, OpenAI::Internal::Util::AnyHash) + file: T.any(OpenAI::Models::Chat::ChatCompletionContentPart::File::File, OpenAI::Internal::AnyHash) ) .void end @@ -28,7 +28,7 @@ module OpenAI # generation. sig do params( - file: T.any(OpenAI::Models::Chat::ChatCompletionContentPart::File::File, OpenAI::Internal::Util::AnyHash), + file: T.any(OpenAI::Models::Chat::ChatCompletionContentPart::File::File, OpenAI::Internal::AnyHash), type: Symbol ) .returns(T.attached_class) @@ -40,7 +40,7 @@ module OpenAI def to_hash end - class File < OpenAI::BaseModel + class File < OpenAI::Internal::Type::BaseModel # The base64 encoded file data, used when passing the file to the model as a # string. sig { returns(T.nilable(String)) } diff --git a/rbi/lib/openai/models/chat/chat_completion_content_part_image.rbi b/rbi/lib/openai/models/chat/chat_completion_content_part_image.rbi index 4ba1cace..3bb8586b 100644 --- a/rbi/lib/openai/models/chat/chat_completion_content_part_image.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_content_part_image.rbi @@ -3,13 +3,13 @@ module OpenAI module Models module Chat - class ChatCompletionContentPartImage < OpenAI::BaseModel + class ChatCompletionContentPartImage < OpenAI::Internal::Type::BaseModel sig { returns(OpenAI::Models::Chat::ChatCompletionContentPartImage::ImageURL) } attr_reader :image_url sig do params( - image_url: T.any(OpenAI::Models::Chat::ChatCompletionContentPartImage::ImageURL, OpenAI::Internal::Util::AnyHash) + image_url: T.any(OpenAI::Models::Chat::ChatCompletionContentPartImage::ImageURL, OpenAI::Internal::AnyHash) ) .void end @@ -22,7 +22,7 @@ module OpenAI # Learn about [image inputs](https://platform.openai.com/docs/guides/vision). sig do params( - image_url: T.any(OpenAI::Models::Chat::ChatCompletionContentPartImage::ImageURL, OpenAI::Internal::Util::AnyHash), + image_url: T.any(OpenAI::Models::Chat::ChatCompletionContentPartImage::ImageURL, OpenAI::Internal::AnyHash), type: Symbol ) .returns(T.attached_class) @@ -36,7 +36,7 @@ module OpenAI def to_hash end - class ImageURL < OpenAI::BaseModel + class ImageURL < OpenAI::Internal::Type::BaseModel # Either a URL of the image or the base64 encoded image data. sig { returns(String) } attr_accessor :url @@ -71,7 +71,7 @@ module OpenAI # Specifies the detail level of the image. Learn more in the # [Vision guide](https://platform.openai.com/docs/guides/vision#low-or-high-fidelity-image-understanding). module Detail - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Chat::ChatCompletionContentPartImage::ImageURL::Detail) } diff --git a/rbi/lib/openai/models/chat/chat_completion_content_part_input_audio.rbi b/rbi/lib/openai/models/chat/chat_completion_content_part_input_audio.rbi index 61ce5857..19fda6e5 100644 --- a/rbi/lib/openai/models/chat/chat_completion_content_part_input_audio.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_content_part_input_audio.rbi @@ -3,16 +3,13 @@ module OpenAI module Models module Chat - class ChatCompletionContentPartInputAudio < OpenAI::BaseModel + class ChatCompletionContentPartInputAudio < OpenAI::Internal::Type::BaseModel sig { returns(OpenAI::Models::Chat::ChatCompletionContentPartInputAudio::InputAudio) } attr_reader :input_audio sig do params( - input_audio: T.any( - OpenAI::Models::Chat::ChatCompletionContentPartInputAudio::InputAudio, - OpenAI::Internal::Util::AnyHash - ) + input_audio: T.any(OpenAI::Models::Chat::ChatCompletionContentPartInputAudio::InputAudio, OpenAI::Internal::AnyHash) ) .void end @@ -25,10 +22,7 @@ module OpenAI # Learn about [audio inputs](https://platform.openai.com/docs/guides/audio). sig do params( - input_audio: T.any( - OpenAI::Models::Chat::ChatCompletionContentPartInputAudio::InputAudio, - OpenAI::Internal::Util::AnyHash - ), + input_audio: T.any(OpenAI::Models::Chat::ChatCompletionContentPartInputAudio::InputAudio, OpenAI::Internal::AnyHash), type: Symbol ) .returns(T.attached_class) @@ -45,7 +39,7 @@ module OpenAI def to_hash end - class InputAudio < OpenAI::BaseModel + class InputAudio < OpenAI::Internal::Type::BaseModel # Base64 encoded audio data. sig { returns(String) } attr_accessor :data @@ -78,7 +72,7 @@ module OpenAI # The format of the encoded audio data. Currently supports "wav" and "mp3". module Format - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Chat::ChatCompletionContentPartInputAudio::InputAudio::Format) } diff --git a/rbi/lib/openai/models/chat/chat_completion_content_part_refusal.rbi b/rbi/lib/openai/models/chat/chat_completion_content_part_refusal.rbi index 788cfecf..7431214c 100644 --- a/rbi/lib/openai/models/chat/chat_completion_content_part_refusal.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_content_part_refusal.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionContentPartRefusal < OpenAI::BaseModel + class ChatCompletionContentPartRefusal < OpenAI::Internal::Type::BaseModel # The refusal message generated by the model. sig { returns(String) } attr_accessor :refusal diff --git a/rbi/lib/openai/models/chat/chat_completion_content_part_text.rbi b/rbi/lib/openai/models/chat/chat_completion_content_part_text.rbi index 1cb1a46c..d9189ad1 100644 --- a/rbi/lib/openai/models/chat/chat_completion_content_part_text.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_content_part_text.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionContentPartText < OpenAI::BaseModel + class ChatCompletionContentPartText < OpenAI::Internal::Type::BaseModel # The text content. sig { returns(String) } attr_accessor :text diff --git a/rbi/lib/openai/models/chat/chat_completion_deleted.rbi b/rbi/lib/openai/models/chat/chat_completion_deleted.rbi index 9affbd35..617a2ea5 100644 --- a/rbi/lib/openai/models/chat/chat_completion_deleted.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_deleted.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionDeleted < OpenAI::BaseModel + class ChatCompletionDeleted < OpenAI::Internal::Type::BaseModel # The ID of the chat completion that was deleted. sig { returns(String) } attr_accessor :id diff --git a/rbi/lib/openai/models/chat/chat_completion_developer_message_param.rbi b/rbi/lib/openai/models/chat/chat_completion_developer_message_param.rbi index ebf50ef3..a026c7f1 100644 --- a/rbi/lib/openai/models/chat/chat_completion_developer_message_param.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_developer_message_param.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionDeveloperMessageParam < OpenAI::BaseModel + class ChatCompletionDeveloperMessageParam < OpenAI::Internal::Type::BaseModel # The contents of the developer message. sig { returns(T.any(String, T::Array[OpenAI::Models::Chat::ChatCompletionContentPartText])) } attr_accessor :content @@ -27,7 +27,7 @@ module OpenAI params( content: T.any( String, - T::Array[T.any(OpenAI::Models::Chat::ChatCompletionContentPartText, OpenAI::Internal::Util::AnyHash)] + T::Array[T.any(OpenAI::Models::Chat::ChatCompletionContentPartText, OpenAI::Internal::AnyHash)] ), name: String, role: Symbol @@ -52,7 +52,7 @@ module OpenAI # The contents of the developer message. module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, T::Array[OpenAI::Models::Chat::ChatCompletionContentPartText]]) } def self.variants @@ -60,7 +60,7 @@ module OpenAI ChatCompletionContentPartTextArray = T.let( - OpenAI::ArrayOf[OpenAI::Models::Chat::ChatCompletionContentPartText], + OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Chat::ChatCompletionContentPartText], OpenAI::Internal::Type::Converter ) end diff --git a/rbi/lib/openai/models/chat/chat_completion_function_call_option.rbi b/rbi/lib/openai/models/chat/chat_completion_function_call_option.rbi index 11ce4c45..45e13985 100644 --- a/rbi/lib/openai/models/chat/chat_completion_function_call_option.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_function_call_option.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionFunctionCallOption < OpenAI::BaseModel + class ChatCompletionFunctionCallOption < OpenAI::Internal::Type::BaseModel # The name of the function to call. sig { returns(String) } attr_accessor :name diff --git a/rbi/lib/openai/models/chat/chat_completion_function_message_param.rbi b/rbi/lib/openai/models/chat/chat_completion_function_message_param.rbi index 43ae6ea4..41e66779 100644 --- a/rbi/lib/openai/models/chat/chat_completion_function_message_param.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_function_message_param.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionFunctionMessageParam < OpenAI::BaseModel + class ChatCompletionFunctionMessageParam < OpenAI::Internal::Type::BaseModel # The contents of the function message. sig { returns(T.nilable(String)) } attr_accessor :content diff --git a/rbi/lib/openai/models/chat/chat_completion_message.rbi b/rbi/lib/openai/models/chat/chat_completion_message.rbi index 10aee914..922e88f6 100644 --- a/rbi/lib/openai/models/chat/chat_completion_message.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_message.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionMessage < OpenAI::BaseModel + class ChatCompletionMessage < OpenAI::Internal::Type::BaseModel # The contents of the message. sig { returns(T.nilable(String)) } attr_accessor :content @@ -23,7 +23,7 @@ module OpenAI sig do params( - annotations: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionMessage::Annotation, OpenAI::Internal::Util::AnyHash)] + annotations: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionMessage::Annotation, OpenAI::Internal::AnyHash)] ) .void end @@ -35,12 +35,7 @@ module OpenAI sig { returns(T.nilable(OpenAI::Models::Chat::ChatCompletionAudio)) } attr_reader :audio - sig do - params( - audio: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionAudio, OpenAI::Internal::Util::AnyHash)) - ) - .void - end + sig { params(audio: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionAudio, OpenAI::Internal::AnyHash))).void } attr_writer :audio # Deprecated and replaced by `tool_calls`. The name and arguments of a function @@ -50,7 +45,7 @@ module OpenAI sig do params( - function_call: T.any(OpenAI::Models::Chat::ChatCompletionMessage::FunctionCall, OpenAI::Internal::Util::AnyHash) + function_call: T.any(OpenAI::Models::Chat::ChatCompletionMessage::FunctionCall, OpenAI::Internal::AnyHash) ) .void end @@ -62,7 +57,7 @@ module OpenAI sig do params( - tool_calls: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionMessageToolCall, OpenAI::Internal::Util::AnyHash)] + tool_calls: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionMessageToolCall, OpenAI::Internal::AnyHash)] ) .void end @@ -73,10 +68,10 @@ module OpenAI params( content: T.nilable(String), refusal: T.nilable(String), - annotations: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionMessage::Annotation, OpenAI::Internal::Util::AnyHash)], - audio: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionAudio, OpenAI::Internal::Util::AnyHash)), - function_call: T.any(OpenAI::Models::Chat::ChatCompletionMessage::FunctionCall, OpenAI::Internal::Util::AnyHash), - tool_calls: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionMessageToolCall, OpenAI::Internal::Util::AnyHash)], + annotations: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionMessage::Annotation, OpenAI::Internal::AnyHash)], + audio: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionAudio, OpenAI::Internal::AnyHash)), + function_call: T.any(OpenAI::Models::Chat::ChatCompletionMessage::FunctionCall, OpenAI::Internal::AnyHash), + tool_calls: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionMessageToolCall, OpenAI::Internal::AnyHash)], role: Symbol ) .returns(T.attached_class) @@ -109,7 +104,7 @@ module OpenAI def to_hash end - class Annotation < OpenAI::BaseModel + class Annotation < OpenAI::Internal::Type::BaseModel # The type of the URL citation. Always `url_citation`. sig { returns(Symbol) } attr_accessor :type @@ -120,10 +115,7 @@ module OpenAI sig do params( - url_citation: T.any( - OpenAI::Models::Chat::ChatCompletionMessage::Annotation::URLCitation, - OpenAI::Internal::Util::AnyHash - ) + url_citation: T.any(OpenAI::Models::Chat::ChatCompletionMessage::Annotation::URLCitation, OpenAI::Internal::AnyHash) ) .void end @@ -132,10 +124,7 @@ module OpenAI # A URL citation when using web search. sig do params( - url_citation: T.any( - OpenAI::Models::Chat::ChatCompletionMessage::Annotation::URLCitation, - OpenAI::Internal::Util::AnyHash - ), + url_citation: T.any(OpenAI::Models::Chat::ChatCompletionMessage::Annotation::URLCitation, OpenAI::Internal::AnyHash), type: Symbol ) .returns(T.attached_class) @@ -152,7 +141,7 @@ module OpenAI def to_hash end - class URLCitation < OpenAI::BaseModel + class URLCitation < OpenAI::Internal::Type::BaseModel # The index of the last character of the URL citation in the message. sig { returns(Integer) } attr_accessor :end_index @@ -187,7 +176,7 @@ module OpenAI end end - class FunctionCall < OpenAI::BaseModel + class FunctionCall < OpenAI::Internal::Type::BaseModel # The arguments to call the function with, as generated by the model in JSON # format. Note that the model does not always generate valid JSON, and may # hallucinate parameters not defined by your function schema. Validate the diff --git a/rbi/lib/openai/models/chat/chat_completion_message_param.rbi b/rbi/lib/openai/models/chat/chat_completion_message_param.rbi index a5792ca5..1a4a2cc0 100644 --- a/rbi/lib/openai/models/chat/chat_completion_message_param.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_message_param.rbi @@ -7,7 +7,7 @@ module OpenAI # messages sent by the user. With o1 models and newer, `developer` messages # replace the previous `system` messages. module ChatCompletionMessageParam - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/chat/chat_completion_message_tool_call.rbi b/rbi/lib/openai/models/chat/chat_completion_message_tool_call.rbi index 496d91cb..4104c3ca 100644 --- a/rbi/lib/openai/models/chat/chat_completion_message_tool_call.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_message_tool_call.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionMessageToolCall < OpenAI::BaseModel + class ChatCompletionMessageToolCall < OpenAI::Internal::Type::BaseModel # The ID of the tool call. sig { returns(String) } attr_accessor :id @@ -14,7 +14,7 @@ module OpenAI sig do params( - function: T.any(OpenAI::Models::Chat::ChatCompletionMessageToolCall::Function, OpenAI::Internal::Util::AnyHash) + function: T.any(OpenAI::Models::Chat::ChatCompletionMessageToolCall::Function, OpenAI::Internal::AnyHash) ) .void end @@ -27,7 +27,7 @@ module OpenAI sig do params( id: String, - function: T.any(OpenAI::Models::Chat::ChatCompletionMessageToolCall::Function, OpenAI::Internal::Util::AnyHash), + function: T.any(OpenAI::Models::Chat::ChatCompletionMessageToolCall::Function, OpenAI::Internal::AnyHash), type: Symbol ) .returns(T.attached_class) @@ -44,7 +44,7 @@ module OpenAI def to_hash end - class Function < OpenAI::BaseModel + class Function < OpenAI::Internal::Type::BaseModel # The arguments to call the function with, as generated by the model in JSON # format. Note that the model does not always generate valid JSON, and may # hallucinate parameters not defined by your function schema. Validate the diff --git a/rbi/lib/openai/models/chat/chat_completion_modality.rbi b/rbi/lib/openai/models/chat/chat_completion_modality.rbi index 785b8db7..5e04a7e7 100644 --- a/rbi/lib/openai/models/chat/chat_completion_modality.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_modality.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Chat module ChatCompletionModality - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Chat::ChatCompletionModality) } OrSymbol = diff --git a/rbi/lib/openai/models/chat/chat_completion_named_tool_choice.rbi b/rbi/lib/openai/models/chat/chat_completion_named_tool_choice.rbi index 6a992cda..566f6c99 100644 --- a/rbi/lib/openai/models/chat/chat_completion_named_tool_choice.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_named_tool_choice.rbi @@ -3,13 +3,13 @@ module OpenAI module Models module Chat - class ChatCompletionNamedToolChoice < OpenAI::BaseModel + class ChatCompletionNamedToolChoice < OpenAI::Internal::Type::BaseModel sig { returns(OpenAI::Models::Chat::ChatCompletionNamedToolChoice::Function) } attr_reader :function sig do params( - function: T.any(OpenAI::Models::Chat::ChatCompletionNamedToolChoice::Function, OpenAI::Internal::Util::AnyHash) + function: T.any(OpenAI::Models::Chat::ChatCompletionNamedToolChoice::Function, OpenAI::Internal::AnyHash) ) .void end @@ -23,7 +23,7 @@ module OpenAI # function. sig do params( - function: T.any(OpenAI::Models::Chat::ChatCompletionNamedToolChoice::Function, OpenAI::Internal::Util::AnyHash), + function: T.any(OpenAI::Models::Chat::ChatCompletionNamedToolChoice::Function, OpenAI::Internal::AnyHash), type: Symbol ) .returns(T.attached_class) @@ -35,7 +35,7 @@ module OpenAI def to_hash end - class Function < OpenAI::BaseModel + class Function < OpenAI::Internal::Type::BaseModel # The name of the function to call. sig { returns(String) } attr_accessor :name diff --git a/rbi/lib/openai/models/chat/chat_completion_prediction_content.rbi b/rbi/lib/openai/models/chat/chat_completion_prediction_content.rbi index ac3a3840..674775a6 100644 --- a/rbi/lib/openai/models/chat/chat_completion_prediction_content.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_prediction_content.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionPredictionContent < OpenAI::BaseModel + class ChatCompletionPredictionContent < OpenAI::Internal::Type::BaseModel # The content that should be matched when generating a model response. If # generated tokens would match this content, the entire model response can be # returned much more quickly. @@ -21,7 +21,7 @@ module OpenAI params( content: T.any( String, - T::Array[T.any(OpenAI::Models::Chat::ChatCompletionContentPartText, OpenAI::Internal::Util::AnyHash)] + T::Array[T.any(OpenAI::Models::Chat::ChatCompletionContentPartText, OpenAI::Internal::AnyHash)] ), type: Symbol ) @@ -43,7 +43,7 @@ module OpenAI # generated tokens would match this content, the entire model response can be # returned much more quickly. module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, T::Array[OpenAI::Models::Chat::ChatCompletionContentPartText]]) } def self.variants @@ -51,7 +51,7 @@ module OpenAI ChatCompletionContentPartTextArray = T.let( - OpenAI::ArrayOf[OpenAI::Models::Chat::ChatCompletionContentPartText], + OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Chat::ChatCompletionContentPartText], OpenAI::Internal::Type::Converter ) end diff --git a/rbi/lib/openai/models/chat/chat_completion_role.rbi b/rbi/lib/openai/models/chat/chat_completion_role.rbi index 61b9c3a5..e75ae72c 100644 --- a/rbi/lib/openai/models/chat/chat_completion_role.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_role.rbi @@ -5,7 +5,7 @@ module OpenAI module Chat # The role of the author of a message module ChatCompletionRole - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Chat::ChatCompletionRole) } OrSymbol = T.type_alias { T.any(Symbol, String, OpenAI::Models::Chat::ChatCompletionRole::TaggedSymbol) } diff --git a/rbi/lib/openai/models/chat/chat_completion_stream_options.rbi b/rbi/lib/openai/models/chat/chat_completion_stream_options.rbi index bb0ee0d9..26683cfa 100644 --- a/rbi/lib/openai/models/chat/chat_completion_stream_options.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_stream_options.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionStreamOptions < OpenAI::BaseModel + class ChatCompletionStreamOptions < OpenAI::Internal::Type::BaseModel # If set, an additional chunk will be streamed before the `data: [DONE]` message. # The `usage` field on this chunk shows the token usage statistics for the entire # request, and the `choices` field will always be an empty array. diff --git a/rbi/lib/openai/models/chat/chat_completion_system_message_param.rbi b/rbi/lib/openai/models/chat/chat_completion_system_message_param.rbi index 6ef22108..bb766e74 100644 --- a/rbi/lib/openai/models/chat/chat_completion_system_message_param.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_system_message_param.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionSystemMessageParam < OpenAI::BaseModel + class ChatCompletionSystemMessageParam < OpenAI::Internal::Type::BaseModel # The contents of the system message. sig { returns(T.any(String, T::Array[OpenAI::Models::Chat::ChatCompletionContentPartText])) } attr_accessor :content @@ -27,7 +27,7 @@ module OpenAI params( content: T.any( String, - T::Array[T.any(OpenAI::Models::Chat::ChatCompletionContentPartText, OpenAI::Internal::Util::AnyHash)] + T::Array[T.any(OpenAI::Models::Chat::ChatCompletionContentPartText, OpenAI::Internal::AnyHash)] ), name: String, role: Symbol @@ -52,7 +52,7 @@ module OpenAI # The contents of the system message. module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, T::Array[OpenAI::Models::Chat::ChatCompletionContentPartText]]) } def self.variants @@ -60,7 +60,7 @@ module OpenAI ChatCompletionContentPartTextArray = T.let( - OpenAI::ArrayOf[OpenAI::Models::Chat::ChatCompletionContentPartText], + OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Chat::ChatCompletionContentPartText], OpenAI::Internal::Type::Converter ) end diff --git a/rbi/lib/openai/models/chat/chat_completion_token_logprob.rbi b/rbi/lib/openai/models/chat/chat_completion_token_logprob.rbi index 131c3a45..17cc3821 100644 --- a/rbi/lib/openai/models/chat/chat_completion_token_logprob.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_token_logprob.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionTokenLogprob < OpenAI::BaseModel + class ChatCompletionTokenLogprob < OpenAI::Internal::Type::BaseModel # The token. sig { returns(String) } attr_accessor :token @@ -32,7 +32,7 @@ module OpenAI token: String, bytes: T.nilable(T::Array[Integer]), logprob: Float, - top_logprobs: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionTokenLogprob::TopLogprob, OpenAI::Internal::Util::AnyHash)] + top_logprobs: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionTokenLogprob::TopLogprob, OpenAI::Internal::AnyHash)] ) .returns(T.attached_class) end @@ -53,7 +53,7 @@ module OpenAI def to_hash end - class TopLogprob < OpenAI::BaseModel + class TopLogprob < OpenAI::Internal::Type::BaseModel # The token. sig { returns(String) } attr_accessor :token diff --git a/rbi/lib/openai/models/chat/chat_completion_tool.rbi b/rbi/lib/openai/models/chat/chat_completion_tool.rbi index dd162d31..e35bbd32 100644 --- a/rbi/lib/openai/models/chat/chat_completion_tool.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_tool.rbi @@ -3,11 +3,11 @@ module OpenAI module Models module Chat - class ChatCompletionTool < OpenAI::BaseModel + class ChatCompletionTool < OpenAI::Internal::Type::BaseModel sig { returns(OpenAI::Models::FunctionDefinition) } attr_reader :function - sig { params(function: T.any(OpenAI::Models::FunctionDefinition, OpenAI::Internal::Util::AnyHash)).void } + sig { params(function: T.any(OpenAI::Models::FunctionDefinition, OpenAI::Internal::AnyHash)).void } attr_writer :function # The type of the tool. Currently, only `function` is supported. @@ -15,7 +15,7 @@ module OpenAI attr_accessor :type sig do - params(function: T.any(OpenAI::Models::FunctionDefinition, OpenAI::Internal::Util::AnyHash), type: Symbol) + params(function: T.any(OpenAI::Models::FunctionDefinition, OpenAI::Internal::AnyHash), type: Symbol) .returns(T.attached_class) end def self.new(function:, type: :function) diff --git a/rbi/lib/openai/models/chat/chat_completion_tool_choice_option.rbi b/rbi/lib/openai/models/chat/chat_completion_tool_choice_option.rbi index 48048594..b739c519 100644 --- a/rbi/lib/openai/models/chat/chat_completion_tool_choice_option.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_tool_choice_option.rbi @@ -13,13 +13,13 @@ module OpenAI # `none` is the default when no tools are present. `auto` is the default if tools # are present. module ChatCompletionToolChoiceOption - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # `none` means the model will not call any tool and instead generates a message. # `auto` means the model can pick between generating a message or calling one or # more tools. `required` means the model must call one or more tools. module Auto - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Chat::ChatCompletionToolChoiceOption::Auto) } OrSymbol = diff --git a/rbi/lib/openai/models/chat/chat_completion_tool_message_param.rbi b/rbi/lib/openai/models/chat/chat_completion_tool_message_param.rbi index e49bdc4a..3c34a0a4 100644 --- a/rbi/lib/openai/models/chat/chat_completion_tool_message_param.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_tool_message_param.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionToolMessageParam < OpenAI::BaseModel + class ChatCompletionToolMessageParam < OpenAI::Internal::Type::BaseModel # The contents of the tool message. sig { returns(T.any(String, T::Array[OpenAI::Models::Chat::ChatCompletionContentPartText])) } attr_accessor :content @@ -20,7 +20,7 @@ module OpenAI params( content: T.any( String, - T::Array[T.any(OpenAI::Models::Chat::ChatCompletionContentPartText, OpenAI::Internal::Util::AnyHash)] + T::Array[T.any(OpenAI::Models::Chat::ChatCompletionContentPartText, OpenAI::Internal::AnyHash)] ), tool_call_id: String, role: Symbol @@ -45,7 +45,7 @@ module OpenAI # The contents of the tool message. module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, T::Array[OpenAI::Models::Chat::ChatCompletionContentPartText]]) } def self.variants @@ -53,7 +53,7 @@ module OpenAI ChatCompletionContentPartTextArray = T.let( - OpenAI::ArrayOf[OpenAI::Models::Chat::ChatCompletionContentPartText], + OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Chat::ChatCompletionContentPartText], OpenAI::Internal::Type::Converter ) end diff --git a/rbi/lib/openai/models/chat/chat_completion_user_message_param.rbi b/rbi/lib/openai/models/chat/chat_completion_user_message_param.rbi index 73d892eb..f4a11939 100644 --- a/rbi/lib/openai/models/chat/chat_completion_user_message_param.rbi +++ b/rbi/lib/openai/models/chat/chat_completion_user_message_param.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class ChatCompletionUserMessageParam < OpenAI::BaseModel + class ChatCompletionUserMessageParam < OpenAI::Internal::Type::BaseModel # The contents of the user message. sig do returns( @@ -43,7 +43,7 @@ module OpenAI T::Array[ T.any( OpenAI::Models::Chat::ChatCompletionContentPartText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Chat::ChatCompletionContentPartImage, OpenAI::Models::Chat::ChatCompletionContentPartInputAudio, OpenAI::Models::Chat::ChatCompletionContentPart::File @@ -83,7 +83,7 @@ module OpenAI # The contents of the user message. module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override @@ -106,7 +106,7 @@ module OpenAI ChatCompletionContentPartArray = T.let( - OpenAI::ArrayOf[union: OpenAI::Models::Chat::ChatCompletionContentPart], + OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Chat::ChatCompletionContentPart], OpenAI::Internal::Type::Converter ) end diff --git a/rbi/lib/openai/models/chat/completion_create_params.rbi b/rbi/lib/openai/models/chat/completion_create_params.rbi index 1d25cd61..cdbe6807 100644 --- a/rbi/lib/openai/models/chat/completion_create_params.rbi +++ b/rbi/lib/openai/models/chat/completion_create_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class CompletionCreateParams < OpenAI::BaseModel + class CompletionCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -44,9 +44,7 @@ module OpenAI attr_reader :audio sig do - params( - audio: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionAudioParam, OpenAI::Internal::Util::AnyHash)) - ) + params(audio: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionAudioParam, OpenAI::Internal::AnyHash))) .void end attr_writer :audio @@ -88,7 +86,7 @@ module OpenAI function_call: T.any( OpenAI::Models::Chat::CompletionCreateParams::FunctionCall::FunctionCallMode::OrSymbol, OpenAI::Models::Chat::ChatCompletionFunctionCallOption, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .void @@ -103,7 +101,7 @@ module OpenAI sig do params( - functions: T::Array[T.any(OpenAI::Models::Chat::CompletionCreateParams::Function, OpenAI::Internal::Util::AnyHash)] + functions: T::Array[T.any(OpenAI::Models::Chat::CompletionCreateParams::Function, OpenAI::Internal::AnyHash)] ) .void end @@ -186,7 +184,7 @@ module OpenAI sig do params( - prediction: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionPredictionContent, OpenAI::Internal::Util::AnyHash)) + prediction: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionPredictionContent, OpenAI::Internal::AnyHash)) ) .void end @@ -234,7 +232,7 @@ module OpenAI params( response_format: T.any( OpenAI::Models::ResponseFormatText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::ResponseFormatJSONSchema, OpenAI::Models::ResponseFormatJSONObject ) @@ -285,7 +283,7 @@ module OpenAI sig do params( - stream_options: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionStreamOptions, OpenAI::Internal::Util::AnyHash)) + stream_options: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionStreamOptions, OpenAI::Internal::AnyHash)) ) .void end @@ -324,7 +322,7 @@ module OpenAI tool_choice: T.any( OpenAI::Models::Chat::ChatCompletionToolChoiceOption::Auto::OrSymbol, OpenAI::Models::Chat::ChatCompletionNamedToolChoice, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .void @@ -337,10 +335,7 @@ module OpenAI sig { returns(T.nilable(T::Array[OpenAI::Models::Chat::ChatCompletionTool])) } attr_reader :tools - sig do - params(tools: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionTool, OpenAI::Internal::Util::AnyHash)]) - .void - end + sig { params(tools: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionTool, OpenAI::Internal::AnyHash)]).void } attr_writer :tools # An integer between 0 and 20 specifying the number of most likely tokens to @@ -374,7 +369,7 @@ module OpenAI sig do params( - web_search_options: T.any(OpenAI::Models::Chat::CompletionCreateParams::WebSearchOptions, OpenAI::Internal::Util::AnyHash) + web_search_options: T.any(OpenAI::Models::Chat::CompletionCreateParams::WebSearchOptions, OpenAI::Internal::AnyHash) ) .void end @@ -385,7 +380,7 @@ module OpenAI messages: T::Array[ T.any( OpenAI::Models::Chat::ChatCompletionDeveloperMessageParam, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Chat::ChatCompletionSystemMessageParam, OpenAI::Models::Chat::ChatCompletionUserMessageParam, OpenAI::Models::Chat::ChatCompletionAssistantMessageParam, @@ -394,14 +389,14 @@ module OpenAI ) ], model: T.any(String, OpenAI::Models::ChatModel::OrSymbol), - audio: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionAudioParam, OpenAI::Internal::Util::AnyHash)), + audio: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionAudioParam, OpenAI::Internal::AnyHash)), frequency_penalty: T.nilable(Float), function_call: T.any( OpenAI::Models::Chat::CompletionCreateParams::FunctionCall::FunctionCallMode::OrSymbol, OpenAI::Models::Chat::ChatCompletionFunctionCallOption, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ), - functions: T::Array[T.any(OpenAI::Models::Chat::CompletionCreateParams::Function, OpenAI::Internal::Util::AnyHash)], + functions: T::Array[T.any(OpenAI::Models::Chat::CompletionCreateParams::Function, OpenAI::Internal::AnyHash)], logit_bias: T.nilable(T::Hash[Symbol, Integer]), logprobs: T.nilable(T::Boolean), max_completion_tokens: T.nilable(Integer), @@ -410,12 +405,12 @@ module OpenAI modalities: T.nilable(T::Array[OpenAI::Models::Chat::CompletionCreateParams::Modality::OrSymbol]), n: T.nilable(Integer), parallel_tool_calls: T::Boolean, - prediction: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionPredictionContent, OpenAI::Internal::Util::AnyHash)), + prediction: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionPredictionContent, OpenAI::Internal::AnyHash)), presence_penalty: T.nilable(Float), reasoning_effort: T.nilable(OpenAI::Models::ReasoningEffort::OrSymbol), response_format: T.any( OpenAI::Models::ResponseFormatText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::ResponseFormatJSONSchema, OpenAI::Models::ResponseFormatJSONObject ), @@ -423,19 +418,19 @@ module OpenAI service_tier: T.nilable(OpenAI::Models::Chat::CompletionCreateParams::ServiceTier::OrSymbol), stop: T.nilable(T.any(String, T::Array[String])), store: T.nilable(T::Boolean), - stream_options: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionStreamOptions, OpenAI::Internal::Util::AnyHash)), + stream_options: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionStreamOptions, OpenAI::Internal::AnyHash)), temperature: T.nilable(Float), tool_choice: T.any( OpenAI::Models::Chat::ChatCompletionToolChoiceOption::Auto::OrSymbol, OpenAI::Models::Chat::ChatCompletionNamedToolChoice, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ), - tools: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionTool, OpenAI::Internal::Util::AnyHash)], + tools: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionTool, OpenAI::Internal::AnyHash)], top_logprobs: T.nilable(Integer), top_p: T.nilable(Float), user: String, - web_search_options: T.any(OpenAI::Models::Chat::CompletionCreateParams::WebSearchOptions, OpenAI::Internal::Util::AnyHash), - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + web_search_options: T.any(OpenAI::Models::Chat::CompletionCreateParams::WebSearchOptions, OpenAI::Internal::AnyHash), + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -540,7 +535,7 @@ module OpenAI # [model guide](https://platform.openai.com/docs/models) to browse and compare # available models. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, OpenAI::Models::ChatModel::OrSymbol]) } def self.variants @@ -562,13 +557,13 @@ module OpenAI # `none` is the default when no functions are present. `auto` is the default if # functions are present. module FunctionCall - extend OpenAI::Union + extend OpenAI::Internal::Type::Union # `none` means the model will not call a function and instead generates a message. # `auto` means the model can pick between generating a message or calling a # function. module FunctionCallMode - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Chat::CompletionCreateParams::FunctionCall::FunctionCallMode) } @@ -606,7 +601,7 @@ module OpenAI end end - class Function < OpenAI::BaseModel + class Function < OpenAI::Internal::Type::BaseModel # The name of the function to be called. Must be a-z, A-Z, 0-9, or contain # underscores and dashes, with a maximum length of 64. sig { returns(String) } @@ -648,7 +643,7 @@ module OpenAI end module Modality - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Chat::CompletionCreateParams::Modality) } OrSymbol = @@ -673,7 +668,7 @@ module OpenAI # ensures the message the model generates is valid JSON. Using `json_schema` is # preferred for models that support it. module ResponseFormat - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override @@ -700,7 +695,7 @@ module OpenAI # When this parameter is set, the response body will include the `service_tier` # utilized. module ServiceTier - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Chat::CompletionCreateParams::ServiceTier) } OrSymbol = @@ -717,16 +712,16 @@ module OpenAI # Up to 4 sequences where the API will stop generating further tokens. The # returned text will not contain the stop sequence. module Stop - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, T::Array[String]]) } def self.variants end - StringArray = T.let(OpenAI::ArrayOf[String], OpenAI::Internal::Type::Converter) + StringArray = T.let(OpenAI::Internal::Type::ArrayOf[String], OpenAI::Internal::Type::Converter) end - class WebSearchOptions < OpenAI::BaseModel + class WebSearchOptions < OpenAI::Internal::Type::BaseModel # High level guidance for the amount of context window space to use for the # search. One of `low`, `medium`, or `high`. `medium` is the default. sig do @@ -753,7 +748,7 @@ module OpenAI user_location: T.nilable( T.any( OpenAI::Models::Chat::CompletionCreateParams::WebSearchOptions::UserLocation, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) ) @@ -770,7 +765,7 @@ module OpenAI user_location: T.nilable( T.any( OpenAI::Models::Chat::CompletionCreateParams::WebSearchOptions::UserLocation, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) ) @@ -794,7 +789,7 @@ module OpenAI # High level guidance for the amount of context window space to use for the # search. One of `low`, `medium`, or `high`. `medium` is the default. module SearchContextSize - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Chat::CompletionCreateParams::WebSearchOptions::SearchContextSize) } @@ -833,7 +828,7 @@ module OpenAI end end - class UserLocation < OpenAI::BaseModel + class UserLocation < OpenAI::Internal::Type::BaseModel # Approximate location parameters for the search. sig { returns(OpenAI::Models::Chat::CompletionCreateParams::WebSearchOptions::UserLocation::Approximate) } attr_reader :approximate @@ -842,7 +837,7 @@ module OpenAI params( approximate: T.any( OpenAI::Models::Chat::CompletionCreateParams::WebSearchOptions::UserLocation::Approximate, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .void @@ -858,7 +853,7 @@ module OpenAI params( approximate: T.any( OpenAI::Models::Chat::CompletionCreateParams::WebSearchOptions::UserLocation::Approximate, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ), type: Symbol ) @@ -879,7 +874,7 @@ module OpenAI def to_hash end - class Approximate < OpenAI::BaseModel + class Approximate < OpenAI::Internal::Type::BaseModel # Free text input for the city of the user, e.g. `San Francisco`. sig { returns(T.nilable(String)) } attr_reader :city diff --git a/rbi/lib/openai/models/chat/completion_delete_params.rbi b/rbi/lib/openai/models/chat/completion_delete_params.rbi index b03964f8..e92e61fa 100644 --- a/rbi/lib/openai/models/chat/completion_delete_params.rbi +++ b/rbi/lib/openai/models/chat/completion_delete_params.rbi @@ -3,13 +3,17 @@ module OpenAI module Models module Chat - class CompletionDeleteParams < OpenAI::BaseModel + class CompletionDeleteParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters sig do - params(request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) - .returns(T.attached_class) + params( + request_options: T.any( + OpenAI::RequestOptions, + OpenAI::Internal::AnyHash + ) + ).returns(T.attached_class) end def self.new(request_options: {}) end diff --git a/rbi/lib/openai/models/chat/completion_list_params.rbi b/rbi/lib/openai/models/chat/completion_list_params.rbi index a9fb2282..1759bb11 100644 --- a/rbi/lib/openai/models/chat/completion_list_params.rbi +++ b/rbi/lib/openai/models/chat/completion_list_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class CompletionListParams < OpenAI::BaseModel + class CompletionListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -49,7 +49,7 @@ module OpenAI metadata: T.nilable(T::Hash[Symbol, String]), model: String, order: OpenAI::Models::Chat::CompletionListParams::Order::OrSymbol, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -75,7 +75,7 @@ module OpenAI # Sort order for Chat Completions by timestamp. Use `asc` for ascending order or # `desc` for descending order. Defaults to `asc`. module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Chat::CompletionListParams::Order) } OrSymbol = diff --git a/rbi/lib/openai/models/chat/completion_retrieve_params.rbi b/rbi/lib/openai/models/chat/completion_retrieve_params.rbi index ec5ec147..921f208a 100644 --- a/rbi/lib/openai/models/chat/completion_retrieve_params.rbi +++ b/rbi/lib/openai/models/chat/completion_retrieve_params.rbi @@ -3,13 +3,17 @@ module OpenAI module Models module Chat - class CompletionRetrieveParams < OpenAI::BaseModel + class CompletionRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters sig do - params(request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) - .returns(T.attached_class) + params( + request_options: T.any( + OpenAI::RequestOptions, + OpenAI::Internal::AnyHash + ) + ).returns(T.attached_class) end def self.new(request_options: {}) end diff --git a/rbi/lib/openai/models/chat/completion_update_params.rbi b/rbi/lib/openai/models/chat/completion_update_params.rbi index d895f848..7a044c5b 100644 --- a/rbi/lib/openai/models/chat/completion_update_params.rbi +++ b/rbi/lib/openai/models/chat/completion_update_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Chat - class CompletionUpdateParams < OpenAI::BaseModel + class CompletionUpdateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -19,7 +19,7 @@ module OpenAI sig do params( metadata: T.nilable(T::Hash[Symbol, String]), - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end diff --git a/rbi/lib/openai/models/chat/completions/message_list_params.rbi b/rbi/lib/openai/models/chat/completions/message_list_params.rbi index 5415b3ca..4f3218d7 100644 --- a/rbi/lib/openai/models/chat/completions/message_list_params.rbi +++ b/rbi/lib/openai/models/chat/completions/message_list_params.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module Chat module Completions - class MessageListParams < OpenAI::BaseModel + class MessageListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -35,7 +35,7 @@ module OpenAI after: String, limit: Integer, order: OpenAI::Models::Chat::Completions::MessageListParams::Order::OrSymbol, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -59,7 +59,7 @@ module OpenAI # Sort order for messages by timestamp. Use `asc` for ascending order or `desc` # for descending order. Defaults to `asc`. module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Chat::Completions::MessageListParams::Order) } diff --git a/rbi/lib/openai/models/chat_model.rbi b/rbi/lib/openai/models/chat_model.rbi index 4494fe86..17dc0d14 100644 --- a/rbi/lib/openai/models/chat_model.rbi +++ b/rbi/lib/openai/models/chat_model.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module ChatModel - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::ChatModel) } OrSymbol = T.type_alias { T.any(Symbol, String, OpenAI::Models::ChatModel::TaggedSymbol) } diff --git a/rbi/lib/openai/models/comparison_filter.rbi b/rbi/lib/openai/models/comparison_filter.rbi index 3549a566..1b06ab72 100644 --- a/rbi/lib/openai/models/comparison_filter.rbi +++ b/rbi/lib/openai/models/comparison_filter.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class ComparisonFilter < OpenAI::BaseModel + class ComparisonFilter < OpenAI::Internal::Type::BaseModel # The key to compare against the value. sig { returns(String) } attr_accessor :key @@ -54,7 +54,7 @@ module OpenAI # - `lt`: less than # - `lte`: less than or equal module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::ComparisonFilter::Type) } OrSymbol = T.type_alias { T.any(Symbol, String, OpenAI::Models::ComparisonFilter::Type::TaggedSymbol) } @@ -74,7 +74,7 @@ module OpenAI # The value to compare against the attribute key; supports string, number, or # boolean types. module Value - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, Float, T::Boolean]) } def self.variants diff --git a/rbi/lib/openai/models/completion.rbi b/rbi/lib/openai/models/completion.rbi index cb4efb9a..98496acd 100644 --- a/rbi/lib/openai/models/completion.rbi +++ b/rbi/lib/openai/models/completion.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class Completion < OpenAI::BaseModel + class Completion < OpenAI::Internal::Type::BaseModel # A unique identifier for the completion. sig { returns(String) } attr_accessor :id @@ -37,7 +37,7 @@ module OpenAI sig { returns(T.nilable(OpenAI::Models::CompletionUsage)) } attr_reader :usage - sig { params(usage: T.any(OpenAI::Models::CompletionUsage, OpenAI::Internal::Util::AnyHash)).void } + sig { params(usage: T.any(OpenAI::Models::CompletionUsage, OpenAI::Internal::AnyHash)).void } attr_writer :usage # Represents a completion response from the API. Note: both the streamed and @@ -45,11 +45,11 @@ module OpenAI sig do params( id: String, - choices: T::Array[T.any(OpenAI::Models::CompletionChoice, OpenAI::Internal::Util::AnyHash)], + choices: T::Array[T.any(OpenAI::Models::CompletionChoice, OpenAI::Internal::AnyHash)], created: Integer, model: String, system_fingerprint: String, - usage: T.any(OpenAI::Models::CompletionUsage, OpenAI::Internal::Util::AnyHash), + usage: T.any(OpenAI::Models::CompletionUsage, OpenAI::Internal::AnyHash), object: Symbol ) .returns(T.attached_class) diff --git a/rbi/lib/openai/models/completion_choice.rbi b/rbi/lib/openai/models/completion_choice.rbi index 516615be..6cd7e70d 100644 --- a/rbi/lib/openai/models/completion_choice.rbi +++ b/rbi/lib/openai/models/completion_choice.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class CompletionChoice < OpenAI::BaseModel + class CompletionChoice < OpenAI::Internal::Type::BaseModel # The reason the model stopped generating tokens. This will be `stop` if the model # hit a natural stop point or a provided stop sequence, `length` if the maximum # number of tokens specified in the request was reached, or `content_filter` if @@ -17,9 +17,7 @@ module OpenAI attr_reader :logprobs sig do - params( - logprobs: T.nilable(T.any(OpenAI::Models::CompletionChoice::Logprobs, OpenAI::Internal::Util::AnyHash)) - ) + params(logprobs: T.nilable(T.any(OpenAI::Models::CompletionChoice::Logprobs, OpenAI::Internal::AnyHash))) .void end attr_writer :logprobs @@ -31,7 +29,7 @@ module OpenAI params( finish_reason: OpenAI::Models::CompletionChoice::FinishReason::OrSymbol, index: Integer, - logprobs: T.nilable(T.any(OpenAI::Models::CompletionChoice::Logprobs, OpenAI::Internal::Util::AnyHash)), + logprobs: T.nilable(T.any(OpenAI::Models::CompletionChoice::Logprobs, OpenAI::Internal::AnyHash)), text: String ) .returns(T.attached_class) @@ -58,7 +56,7 @@ module OpenAI # number of tokens specified in the request was reached, or `content_filter` if # content was omitted due to a flag from our content filters. module FinishReason - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::CompletionChoice::FinishReason) } OrSymbol = @@ -73,7 +71,7 @@ module OpenAI end end - class Logprobs < OpenAI::BaseModel + class Logprobs < OpenAI::Internal::Type::BaseModel sig { returns(T.nilable(T::Array[Integer])) } attr_reader :text_offset diff --git a/rbi/lib/openai/models/completion_create_params.rbi b/rbi/lib/openai/models/completion_create_params.rbi index b71c3965..26802234 100644 --- a/rbi/lib/openai/models/completion_create_params.rbi +++ b/rbi/lib/openai/models/completion_create_params.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class CompletionCreateParams < OpenAI::BaseModel + class CompletionCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -119,7 +119,7 @@ module OpenAI sig do params( - stream_options: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionStreamOptions, OpenAI::Internal::Util::AnyHash)) + stream_options: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionStreamOptions, OpenAI::Internal::AnyHash)) ) .void end @@ -177,12 +177,12 @@ module OpenAI presence_penalty: T.nilable(Float), seed: T.nilable(Integer), stop: T.nilable(T.any(String, T::Array[String])), - stream_options: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionStreamOptions, OpenAI::Internal::Util::AnyHash)), + stream_options: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionStreamOptions, OpenAI::Internal::AnyHash)), suffix: T.nilable(String), temperature: T.nilable(Float), top_p: T.nilable(Float), user: String, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -249,7 +249,7 @@ module OpenAI # [Model overview](https://platform.openai.com/docs/models) for descriptions of # them. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, OpenAI::Models::CompletionCreateParams::Model::OrSymbol]) } def self.variants @@ -272,32 +272,33 @@ module OpenAI # training, so if a prompt is not specified the model will generate as if from the # beginning of a new document. module Prompt - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, T::Array[String], T::Array[Integer], T::Array[T::Array[Integer]]]) } def self.variants end - StringArray = T.let(OpenAI::ArrayOf[String], OpenAI::Internal::Type::Converter) + StringArray = T.let(OpenAI::Internal::Type::ArrayOf[String], OpenAI::Internal::Type::Converter) - IntegerArray = T.let(OpenAI::ArrayOf[Integer], OpenAI::Internal::Type::Converter) + IntegerArray = T.let(OpenAI::Internal::Type::ArrayOf[Integer], OpenAI::Internal::Type::Converter) - ArrayOfToken2DArray = T.let( - OpenAI::ArrayOf[OpenAI::ArrayOf[Integer]], - OpenAI::Internal::Type::Converter - ) + ArrayOfToken2DArray = + T.let( + OpenAI::Internal::Type::ArrayOf[OpenAI::Internal::Type::ArrayOf[Integer]], + OpenAI::Internal::Type::Converter + ) end # Up to 4 sequences where the API will stop generating further tokens. The # returned text will not contain the stop sequence. module Stop - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, T::Array[String]]) } def self.variants end - StringArray = T.let(OpenAI::ArrayOf[String], OpenAI::Internal::Type::Converter) + StringArray = T.let(OpenAI::Internal::Type::ArrayOf[String], OpenAI::Internal::Type::Converter) end end end diff --git a/rbi/lib/openai/models/completion_usage.rbi b/rbi/lib/openai/models/completion_usage.rbi index f059d80b..fb3951e4 100644 --- a/rbi/lib/openai/models/completion_usage.rbi +++ b/rbi/lib/openai/models/completion_usage.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class CompletionUsage < OpenAI::BaseModel + class CompletionUsage < OpenAI::Internal::Type::BaseModel # Number of tokens in the generated completion. sig { returns(Integer) } attr_accessor :completion_tokens @@ -21,7 +21,7 @@ module OpenAI sig do params( - completion_tokens_details: T.any(OpenAI::Models::CompletionUsage::CompletionTokensDetails, OpenAI::Internal::Util::AnyHash) + completion_tokens_details: T.any(OpenAI::Models::CompletionUsage::CompletionTokensDetails, OpenAI::Internal::AnyHash) ) .void end @@ -33,7 +33,7 @@ module OpenAI sig do params( - prompt_tokens_details: T.any(OpenAI::Models::CompletionUsage::PromptTokensDetails, OpenAI::Internal::Util::AnyHash) + prompt_tokens_details: T.any(OpenAI::Models::CompletionUsage::PromptTokensDetails, OpenAI::Internal::AnyHash) ) .void end @@ -45,8 +45,8 @@ module OpenAI completion_tokens: Integer, prompt_tokens: Integer, total_tokens: Integer, - completion_tokens_details: T.any(OpenAI::Models::CompletionUsage::CompletionTokensDetails, OpenAI::Internal::Util::AnyHash), - prompt_tokens_details: T.any(OpenAI::Models::CompletionUsage::PromptTokensDetails, OpenAI::Internal::Util::AnyHash) + completion_tokens_details: T.any(OpenAI::Models::CompletionUsage::CompletionTokensDetails, OpenAI::Internal::AnyHash), + prompt_tokens_details: T.any(OpenAI::Models::CompletionUsage::PromptTokensDetails, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -74,7 +74,7 @@ module OpenAI def to_hash end - class CompletionTokensDetails < OpenAI::BaseModel + class CompletionTokensDetails < OpenAI::Internal::Type::BaseModel # When using Predicted Outputs, the number of tokens in the prediction that # appeared in the completion. sig { returns(T.nilable(Integer)) } @@ -140,7 +140,7 @@ module OpenAI end end - class PromptTokensDetails < OpenAI::BaseModel + class PromptTokensDetails < OpenAI::Internal::Type::BaseModel # Audio input tokens present in the prompt. sig { returns(T.nilable(Integer)) } attr_reader :audio_tokens diff --git a/rbi/lib/openai/models/compound_filter.rbi b/rbi/lib/openai/models/compound_filter.rbi index 7377b320..056c00fe 100644 --- a/rbi/lib/openai/models/compound_filter.rbi +++ b/rbi/lib/openai/models/compound_filter.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class CompoundFilter < OpenAI::BaseModel + class CompoundFilter < OpenAI::Internal::Type::BaseModel # Array of filters to combine. Items can be `ComparisonFilter` or # `CompoundFilter`. sig { returns(T::Array[T.any(OpenAI::Models::ComparisonFilter, T.anything)]) } @@ -15,7 +15,7 @@ module OpenAI # Combine multiple filters using `and` or `or`. sig do params( - filters: T::Array[T.any(OpenAI::Models::ComparisonFilter, OpenAI::Internal::Util::AnyHash, T.anything)], + filters: T::Array[T.any(OpenAI::Models::ComparisonFilter, OpenAI::Internal::AnyHash, T.anything)], type: OpenAI::Models::CompoundFilter::Type::OrSymbol ) .returns(T.attached_class) @@ -38,7 +38,7 @@ module OpenAI # A filter used to compare a specified attribute key to a given value using a # defined comparison operation. module Filter - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([OpenAI::Models::ComparisonFilter, T.anything]) } def self.variants @@ -47,7 +47,7 @@ module OpenAI # Type of operation: `and` or `or`. module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::CompoundFilter::Type) } OrSymbol = T.type_alias { T.any(Symbol, String, OpenAI::Models::CompoundFilter::Type::TaggedSymbol) } diff --git a/rbi/lib/openai/models/create_embedding_response.rbi b/rbi/lib/openai/models/create_embedding_response.rbi index d970b614..e76ac6c3 100644 --- a/rbi/lib/openai/models/create_embedding_response.rbi +++ b/rbi/lib/openai/models/create_embedding_response.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class CreateEmbeddingResponse < OpenAI::BaseModel + class CreateEmbeddingResponse < OpenAI::Internal::Type::BaseModel # The list of embeddings generated by the model. sig { returns(T::Array[OpenAI::Models::Embedding]) } attr_accessor :data @@ -19,14 +19,14 @@ module OpenAI sig { returns(OpenAI::Models::CreateEmbeddingResponse::Usage) } attr_reader :usage - sig { params(usage: T.any(OpenAI::Models::CreateEmbeddingResponse::Usage, OpenAI::Internal::Util::AnyHash)).void } + sig { params(usage: T.any(OpenAI::Models::CreateEmbeddingResponse::Usage, OpenAI::Internal::AnyHash)).void } attr_writer :usage sig do params( - data: T::Array[T.any(OpenAI::Models::Embedding, OpenAI::Internal::Util::AnyHash)], + data: T::Array[T.any(OpenAI::Models::Embedding, OpenAI::Internal::AnyHash)], model: String, - usage: T.any(OpenAI::Models::CreateEmbeddingResponse::Usage, OpenAI::Internal::Util::AnyHash), + usage: T.any(OpenAI::Models::CreateEmbeddingResponse::Usage, OpenAI::Internal::AnyHash), object: Symbol ) .returns(T.attached_class) @@ -48,7 +48,7 @@ module OpenAI def to_hash end - class Usage < OpenAI::BaseModel + class Usage < OpenAI::Internal::Type::BaseModel # The number of tokens used by the prompt. sig { returns(Integer) } attr_accessor :prompt_tokens diff --git a/rbi/lib/openai/models/embedding.rbi b/rbi/lib/openai/models/embedding.rbi index 90ad0f45..1f6e9020 100644 --- a/rbi/lib/openai/models/embedding.rbi +++ b/rbi/lib/openai/models/embedding.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class Embedding < OpenAI::BaseModel + class Embedding < OpenAI::Internal::Type::BaseModel # The embedding vector, which is a list of floats. The length of vector depends on # the model as listed in the # [embedding guide](https://platform.openai.com/docs/guides/embeddings). diff --git a/rbi/lib/openai/models/embedding_create_params.rbi b/rbi/lib/openai/models/embedding_create_params.rbi index 47c96aeb..5d8acf9b 100644 --- a/rbi/lib/openai/models/embedding_create_params.rbi +++ b/rbi/lib/openai/models/embedding_create_params.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class EmbeddingCreateParams < OpenAI::BaseModel + class EmbeddingCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -57,7 +57,7 @@ module OpenAI dimensions: Integer, encoding_format: OpenAI::Models::EmbeddingCreateParams::EncodingFormat::OrSymbol, user: String, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -89,20 +89,21 @@ module OpenAI # for counting tokens. Some models may also impose a limit on total number of # tokens summed across inputs. module Input - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, T::Array[String], T::Array[Integer], T::Array[T::Array[Integer]]]) } def self.variants end - StringArray = T.let(OpenAI::ArrayOf[String], OpenAI::Internal::Type::Converter) + StringArray = T.let(OpenAI::Internal::Type::ArrayOf[String], OpenAI::Internal::Type::Converter) - IntegerArray = T.let(OpenAI::ArrayOf[Integer], OpenAI::Internal::Type::Converter) + IntegerArray = T.let(OpenAI::Internal::Type::ArrayOf[Integer], OpenAI::Internal::Type::Converter) - ArrayOfToken2DArray = T.let( - OpenAI::ArrayOf[OpenAI::ArrayOf[Integer]], - OpenAI::Internal::Type::Converter - ) + ArrayOfToken2DArray = + T.let( + OpenAI::Internal::Type::ArrayOf[OpenAI::Internal::Type::ArrayOf[Integer]], + OpenAI::Internal::Type::Converter + ) end # ID of the model to use. You can use the @@ -111,7 +112,7 @@ module OpenAI # [Model overview](https://platform.openai.com/docs/models) for descriptions of # them. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, OpenAI::Models::EmbeddingModel::OrSymbol]) } def self.variants @@ -121,7 +122,7 @@ module OpenAI # The format to return the embeddings in. Can be either `float` or # [`base64`](https://pypi.org/project/pybase64/). module EncodingFormat - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::EmbeddingCreateParams::EncodingFormat) } OrSymbol = diff --git a/rbi/lib/openai/models/embedding_model.rbi b/rbi/lib/openai/models/embedding_model.rbi index 94781b12..fb2fd9ed 100644 --- a/rbi/lib/openai/models/embedding_model.rbi +++ b/rbi/lib/openai/models/embedding_model.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module EmbeddingModel - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::EmbeddingModel) } OrSymbol = T.type_alias { T.any(Symbol, String, OpenAI::Models::EmbeddingModel::TaggedSymbol) } diff --git a/rbi/lib/openai/models/error_object.rbi b/rbi/lib/openai/models/error_object.rbi index 134b6d66..dfea9b56 100644 --- a/rbi/lib/openai/models/error_object.rbi +++ b/rbi/lib/openai/models/error_object.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class ErrorObject < OpenAI::BaseModel + class ErrorObject < OpenAI::Internal::Type::BaseModel sig { returns(T.nilable(String)) } attr_accessor :code diff --git a/rbi/lib/openai/models/file_chunking_strategy.rbi b/rbi/lib/openai/models/file_chunking_strategy.rbi index 4e9e85c9..8e896180 100644 --- a/rbi/lib/openai/models/file_chunking_strategy.rbi +++ b/rbi/lib/openai/models/file_chunking_strategy.rbi @@ -4,7 +4,7 @@ module OpenAI module Models # The strategy used to chunk the file. module FileChunkingStrategy - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/file_chunking_strategy_param.rbi b/rbi/lib/openai/models/file_chunking_strategy_param.rbi index 17f0755f..9f69a517 100644 --- a/rbi/lib/openai/models/file_chunking_strategy_param.rbi +++ b/rbi/lib/openai/models/file_chunking_strategy_param.rbi @@ -5,7 +5,7 @@ module OpenAI # The chunking strategy used to chunk the file(s). If not set, will use the `auto` # strategy. Only applicable if `file_ids` is non-empty. module FileChunkingStrategyParam - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/file_content_params.rbi b/rbi/lib/openai/models/file_content_params.rbi index 773bdb83..f753c7d2 100644 --- a/rbi/lib/openai/models/file_content_params.rbi +++ b/rbi/lib/openai/models/file_content_params.rbi @@ -2,13 +2,17 @@ module OpenAI module Models - class FileContentParams < OpenAI::BaseModel + class FileContentParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters sig do - params(request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) - .returns(T.attached_class) + params( + request_options: T.any( + OpenAI::RequestOptions, + OpenAI::Internal::AnyHash + ) + ).returns(T.attached_class) end def self.new(request_options: {}) end diff --git a/rbi/lib/openai/models/file_create_params.rbi b/rbi/lib/openai/models/file_create_params.rbi index f665b3dc..c2a7e47f 100644 --- a/rbi/lib/openai/models/file_create_params.rbi +++ b/rbi/lib/openai/models/file_create_params.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class FileCreateParams < OpenAI::BaseModel + class FileCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -21,7 +21,7 @@ module OpenAI params( file: T.any(IO, StringIO), purpose: OpenAI::Models::FilePurpose::OrSymbol, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end diff --git a/rbi/lib/openai/models/file_delete_params.rbi b/rbi/lib/openai/models/file_delete_params.rbi index 354f6382..af5aa65f 100644 --- a/rbi/lib/openai/models/file_delete_params.rbi +++ b/rbi/lib/openai/models/file_delete_params.rbi @@ -2,13 +2,17 @@ module OpenAI module Models - class FileDeleteParams < OpenAI::BaseModel + class FileDeleteParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters sig do - params(request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) - .returns(T.attached_class) + params( + request_options: T.any( + OpenAI::RequestOptions, + OpenAI::Internal::AnyHash + ) + ).returns(T.attached_class) end def self.new(request_options: {}) end diff --git a/rbi/lib/openai/models/file_deleted.rbi b/rbi/lib/openai/models/file_deleted.rbi index fd6fac2b..6b362f3b 100644 --- a/rbi/lib/openai/models/file_deleted.rbi +++ b/rbi/lib/openai/models/file_deleted.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class FileDeleted < OpenAI::BaseModel + class FileDeleted < OpenAI::Internal::Type::BaseModel sig { returns(String) } attr_accessor :id diff --git a/rbi/lib/openai/models/file_list_params.rbi b/rbi/lib/openai/models/file_list_params.rbi index c1cc0765..035e32d3 100644 --- a/rbi/lib/openai/models/file_list_params.rbi +++ b/rbi/lib/openai/models/file_list_params.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class FileListParams < OpenAI::BaseModel + class FileListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -45,7 +45,7 @@ module OpenAI limit: Integer, order: OpenAI::Models::FileListParams::Order::OrSymbol, purpose: String, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -70,7 +70,7 @@ module OpenAI # Sort order by the `created_at` timestamp of the objects. `asc` for ascending # order and `desc` for descending order. module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::FileListParams::Order) } OrSymbol = T.type_alias { T.any(Symbol, String, OpenAI::Models::FileListParams::Order::TaggedSymbol) } diff --git a/rbi/lib/openai/models/file_object.rbi b/rbi/lib/openai/models/file_object.rbi index e8452219..c7f26a14 100644 --- a/rbi/lib/openai/models/file_object.rbi +++ b/rbi/lib/openai/models/file_object.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class FileObject < OpenAI::BaseModel + class FileObject < OpenAI::Internal::Type::BaseModel # The file identifier, which can be referenced in the API endpoints. sig { returns(String) } attr_accessor :id @@ -100,7 +100,7 @@ module OpenAI # `assistants_output`, `batch`, `batch_output`, `fine-tune`, `fine-tune-results` # and `vision`. module Purpose - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::FileObject::Purpose) } OrSymbol = T.type_alias { T.any(Symbol, String, OpenAI::Models::FileObject::Purpose::TaggedSymbol) } @@ -121,7 +121,7 @@ module OpenAI # Deprecated. The current status of the file, which can be either `uploaded`, # `processed`, or `error`. module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::FileObject::Status) } OrSymbol = T.type_alias { T.any(Symbol, String, OpenAI::Models::FileObject::Status::TaggedSymbol) } diff --git a/rbi/lib/openai/models/file_purpose.rbi b/rbi/lib/openai/models/file_purpose.rbi index 1f7bb4f9..5514076c 100644 --- a/rbi/lib/openai/models/file_purpose.rbi +++ b/rbi/lib/openai/models/file_purpose.rbi @@ -7,7 +7,7 @@ module OpenAI # fine-tuning - `vision`: Images used for vision fine-tuning - `user_data`: # Flexible file type for any purpose - `evals`: Used for eval data sets module FilePurpose - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::FilePurpose) } OrSymbol = T.type_alias { T.any(Symbol, String, OpenAI::Models::FilePurpose::TaggedSymbol) } diff --git a/rbi/lib/openai/models/file_retrieve_params.rbi b/rbi/lib/openai/models/file_retrieve_params.rbi index d34e1631..25676e12 100644 --- a/rbi/lib/openai/models/file_retrieve_params.rbi +++ b/rbi/lib/openai/models/file_retrieve_params.rbi @@ -2,13 +2,17 @@ module OpenAI module Models - class FileRetrieveParams < OpenAI::BaseModel + class FileRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters sig do - params(request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) - .returns(T.attached_class) + params( + request_options: T.any( + OpenAI::RequestOptions, + OpenAI::Internal::AnyHash + ) + ).returns(T.attached_class) end def self.new(request_options: {}) end diff --git a/rbi/lib/openai/models/fine_tuning/fine_tuning_job.rbi b/rbi/lib/openai/models/fine_tuning/fine_tuning_job.rbi index c6541e22..810bf8fd 100644 --- a/rbi/lib/openai/models/fine_tuning/fine_tuning_job.rbi +++ b/rbi/lib/openai/models/fine_tuning/fine_tuning_job.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module FineTuning - class FineTuningJob < OpenAI::BaseModel + class FineTuningJob < OpenAI::Internal::Type::BaseModel # The object identifier, which can be referenced in the API endpoints. sig { returns(String) } attr_accessor :id @@ -19,7 +19,7 @@ module OpenAI sig do params( - error: T.nilable(T.any(OpenAI::Models::FineTuning::FineTuningJob::Error, OpenAI::Internal::Util::AnyHash)) + error: T.nilable(T.any(OpenAI::Models::FineTuning::FineTuningJob::Error, OpenAI::Internal::AnyHash)) ) .void end @@ -42,7 +42,7 @@ module OpenAI sig do params( - hyperparameters: T.any(OpenAI::Models::FineTuning::FineTuningJob::Hyperparameters, OpenAI::Internal::Util::AnyHash) + hyperparameters: T.any(OpenAI::Models::FineTuning::FineTuningJob::Hyperparameters, OpenAI::Internal::AnyHash) ) .void end @@ -113,10 +113,7 @@ module OpenAI sig { returns(T.nilable(OpenAI::Models::FineTuning::FineTuningJob::Method)) } attr_reader :method_ - sig do - params(method_: T.any(OpenAI::Models::FineTuning::FineTuningJob::Method, OpenAI::Internal::Util::AnyHash)) - .void - end + sig { params(method_: T.any(OpenAI::Models::FineTuning::FineTuningJob::Method, OpenAI::Internal::AnyHash)).void } attr_writer :method_ # The `fine_tuning.job` object represents a fine-tuning job that has been created @@ -125,10 +122,10 @@ module OpenAI params( id: String, created_at: Integer, - error: T.nilable(T.any(OpenAI::Models::FineTuning::FineTuningJob::Error, OpenAI::Internal::Util::AnyHash)), + error: T.nilable(T.any(OpenAI::Models::FineTuning::FineTuningJob::Error, OpenAI::Internal::AnyHash)), fine_tuned_model: T.nilable(String), finished_at: T.nilable(Integer), - hyperparameters: T.any(OpenAI::Models::FineTuning::FineTuningJob::Hyperparameters, OpenAI::Internal::Util::AnyHash), + hyperparameters: T.any(OpenAI::Models::FineTuning::FineTuningJob::Hyperparameters, OpenAI::Internal::AnyHash), model: String, organization_id: String, result_files: T::Array[String], @@ -139,10 +136,10 @@ module OpenAI validation_file: T.nilable(String), estimated_finish: T.nilable(Integer), integrations: T.nilable( - T::Array[T.any(OpenAI::Models::FineTuning::FineTuningJobWandbIntegrationObject, OpenAI::Internal::Util::AnyHash)] + T::Array[T.any(OpenAI::Models::FineTuning::FineTuningJobWandbIntegrationObject, OpenAI::Internal::AnyHash)] ), metadata: T.nilable(T::Hash[Symbol, String]), - method_: T.any(OpenAI::Models::FineTuning::FineTuningJob::Method, OpenAI::Internal::Util::AnyHash), + method_: T.any(OpenAI::Models::FineTuning::FineTuningJob::Method, OpenAI::Internal::AnyHash), object: Symbol ) .returns(T.attached_class) @@ -199,7 +196,7 @@ module OpenAI def to_hash end - class Error < OpenAI::BaseModel + class Error < OpenAI::Internal::Type::BaseModel # A machine-readable error code. sig { returns(String) } attr_accessor :code @@ -224,7 +221,7 @@ module OpenAI end end - class Hyperparameters < OpenAI::BaseModel + class Hyperparameters < OpenAI::Internal::Type::BaseModel # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. sig { returns(T.nilable(T.any(Symbol, Integer))) } @@ -278,7 +275,7 @@ module OpenAI # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. module BatchSize - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([Symbol, Integer]) } def self.variants @@ -288,7 +285,7 @@ module OpenAI # Scaling factor for the learning rate. A smaller learning rate may be useful to # avoid overfitting. module LearningRateMultiplier - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([Symbol, Float]) } def self.variants @@ -298,7 +295,7 @@ module OpenAI # The number of epochs to train the model for. An epoch refers to one full cycle # through the training dataset. module NEpochs - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([Symbol, Integer]) } def self.variants @@ -309,7 +306,7 @@ module OpenAI # The current status of the fine-tuning job, which can be either # `validating_files`, `queued`, `running`, `succeeded`, `failed`, or `cancelled`. module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::FineTuning::FineTuningJob::Status) } OrSymbol = @@ -328,17 +325,12 @@ module OpenAI end end - class Method < OpenAI::BaseModel + class Method < OpenAI::Internal::Type::BaseModel # Configuration for the DPO fine-tuning method. sig { returns(T.nilable(OpenAI::Models::FineTuning::FineTuningJob::Method::Dpo)) } attr_reader :dpo - sig do - params( - dpo: T.any(OpenAI::Models::FineTuning::FineTuningJob::Method::Dpo, OpenAI::Internal::Util::AnyHash) - ) - .void - end + sig { params(dpo: T.any(OpenAI::Models::FineTuning::FineTuningJob::Method::Dpo, OpenAI::Internal::AnyHash)).void } attr_writer :dpo # Configuration for the supervised fine-tuning method. @@ -347,7 +339,7 @@ module OpenAI sig do params( - supervised: T.any(OpenAI::Models::FineTuning::FineTuningJob::Method::Supervised, OpenAI::Internal::Util::AnyHash) + supervised: T.any(OpenAI::Models::FineTuning::FineTuningJob::Method::Supervised, OpenAI::Internal::AnyHash) ) .void end @@ -363,8 +355,8 @@ module OpenAI # The method used for fine-tuning. sig do params( - dpo: T.any(OpenAI::Models::FineTuning::FineTuningJob::Method::Dpo, OpenAI::Internal::Util::AnyHash), - supervised: T.any(OpenAI::Models::FineTuning::FineTuningJob::Method::Supervised, OpenAI::Internal::Util::AnyHash), + dpo: T.any(OpenAI::Models::FineTuning::FineTuningJob::Method::Dpo, OpenAI::Internal::AnyHash), + supervised: T.any(OpenAI::Models::FineTuning::FineTuningJob::Method::Supervised, OpenAI::Internal::AnyHash), type: OpenAI::Models::FineTuning::FineTuningJob::Method::Type::OrSymbol ) .returns(T.attached_class) @@ -385,17 +377,14 @@ module OpenAI def to_hash end - class Dpo < OpenAI::BaseModel + class Dpo < OpenAI::Internal::Type::BaseModel # The hyperparameters used for the fine-tuning job. sig { returns(T.nilable(OpenAI::Models::FineTuning::FineTuningJob::Method::Dpo::Hyperparameters)) } attr_reader :hyperparameters sig do params( - hyperparameters: T.any( - OpenAI::Models::FineTuning::FineTuningJob::Method::Dpo::Hyperparameters, - OpenAI::Internal::Util::AnyHash - ) + hyperparameters: T.any(OpenAI::Models::FineTuning::FineTuningJob::Method::Dpo::Hyperparameters, OpenAI::Internal::AnyHash) ) .void end @@ -404,10 +393,7 @@ module OpenAI # Configuration for the DPO fine-tuning method. sig do params( - hyperparameters: T.any( - OpenAI::Models::FineTuning::FineTuningJob::Method::Dpo::Hyperparameters, - OpenAI::Internal::Util::AnyHash - ) + hyperparameters: T.any(OpenAI::Models::FineTuning::FineTuningJob::Method::Dpo::Hyperparameters, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -421,7 +407,7 @@ module OpenAI def to_hash end - class Hyperparameters < OpenAI::BaseModel + class Hyperparameters < OpenAI::Internal::Type::BaseModel # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. sig { returns(T.nilable(T.any(Symbol, Integer))) } @@ -484,7 +470,7 @@ module OpenAI # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. module BatchSize - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([Symbol, Integer]) } def self.variants @@ -494,7 +480,7 @@ module OpenAI # The beta value for the DPO method. A higher beta value will increase the weight # of the penalty between the policy and reference model. module Beta - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([Symbol, Float]) } def self.variants @@ -504,7 +490,7 @@ module OpenAI # Scaling factor for the learning rate. A smaller learning rate may be useful to # avoid overfitting. module LearningRateMultiplier - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([Symbol, Float]) } def self.variants @@ -514,7 +500,7 @@ module OpenAI # The number of epochs to train the model for. An epoch refers to one full cycle # through the training dataset. module NEpochs - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([Symbol, Integer]) } def self.variants @@ -523,7 +509,7 @@ module OpenAI end end - class Supervised < OpenAI::BaseModel + class Supervised < OpenAI::Internal::Type::BaseModel # The hyperparameters used for the fine-tuning job. sig { returns(T.nilable(OpenAI::Models::FineTuning::FineTuningJob::Method::Supervised::Hyperparameters)) } attr_reader :hyperparameters @@ -532,7 +518,7 @@ module OpenAI params( hyperparameters: T.any( OpenAI::Models::FineTuning::FineTuningJob::Method::Supervised::Hyperparameters, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .void @@ -544,7 +530,7 @@ module OpenAI params( hyperparameters: T.any( OpenAI::Models::FineTuning::FineTuningJob::Method::Supervised::Hyperparameters, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .returns(T.attached_class) @@ -559,7 +545,7 @@ module OpenAI def to_hash end - class Hyperparameters < OpenAI::BaseModel + class Hyperparameters < OpenAI::Internal::Type::BaseModel # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. sig { returns(T.nilable(T.any(Symbol, Integer))) } @@ -612,7 +598,7 @@ module OpenAI # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. module BatchSize - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([Symbol, Integer]) } def self.variants @@ -622,7 +608,7 @@ module OpenAI # Scaling factor for the learning rate. A smaller learning rate may be useful to # avoid overfitting. module LearningRateMultiplier - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([Symbol, Float]) } def self.variants @@ -632,7 +618,7 @@ module OpenAI # The number of epochs to train the model for. An epoch refers to one full cycle # through the training dataset. module NEpochs - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([Symbol, Integer]) } def self.variants @@ -643,7 +629,7 @@ module OpenAI # The type of method. Is either `supervised` or `dpo`. module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::FineTuning::FineTuningJob::Method::Type) } OrSymbol = diff --git a/rbi/lib/openai/models/fine_tuning/fine_tuning_job_event.rbi b/rbi/lib/openai/models/fine_tuning/fine_tuning_job_event.rbi index 811f8518..f9a16f3d 100644 --- a/rbi/lib/openai/models/fine_tuning/fine_tuning_job_event.rbi +++ b/rbi/lib/openai/models/fine_tuning/fine_tuning_job_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module FineTuning - class FineTuningJobEvent < OpenAI::BaseModel + class FineTuningJobEvent < OpenAI::Internal::Type::BaseModel # The object identifier. sig { returns(String) } attr_accessor :id @@ -81,7 +81,7 @@ module OpenAI # The log level of the event. module Level - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::FineTuning::FineTuningJobEvent::Level) } OrSymbol = @@ -98,7 +98,7 @@ module OpenAI # The type of event. module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::FineTuning::FineTuningJobEvent::Type) } OrSymbol = diff --git a/rbi/lib/openai/models/fine_tuning/fine_tuning_job_wandb_integration.rbi b/rbi/lib/openai/models/fine_tuning/fine_tuning_job_wandb_integration.rbi index 315342d0..561dd861 100644 --- a/rbi/lib/openai/models/fine_tuning/fine_tuning_job_wandb_integration.rbi +++ b/rbi/lib/openai/models/fine_tuning/fine_tuning_job_wandb_integration.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module FineTuning - class FineTuningJobWandbIntegration < OpenAI::BaseModel + class FineTuningJobWandbIntegration < OpenAI::Internal::Type::BaseModel # The name of the project that the new run will be created under. sig { returns(String) } attr_accessor :project diff --git a/rbi/lib/openai/models/fine_tuning/fine_tuning_job_wandb_integration_object.rbi b/rbi/lib/openai/models/fine_tuning/fine_tuning_job_wandb_integration_object.rbi index 78bec198..bb2e26b6 100644 --- a/rbi/lib/openai/models/fine_tuning/fine_tuning_job_wandb_integration_object.rbi +++ b/rbi/lib/openai/models/fine_tuning/fine_tuning_job_wandb_integration_object.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module FineTuning - class FineTuningJobWandbIntegrationObject < OpenAI::BaseModel + class FineTuningJobWandbIntegrationObject < OpenAI::Internal::Type::BaseModel # The type of the integration being enabled for the fine-tuning job sig { returns(Symbol) } attr_accessor :type @@ -16,16 +16,14 @@ module OpenAI attr_reader :wandb sig do - params( - wandb: T.any(OpenAI::Models::FineTuning::FineTuningJobWandbIntegration, OpenAI::Internal::Util::AnyHash) - ) + params(wandb: T.any(OpenAI::Models::FineTuning::FineTuningJobWandbIntegration, OpenAI::Internal::AnyHash)) .void end attr_writer :wandb sig do params( - wandb: T.any(OpenAI::Models::FineTuning::FineTuningJobWandbIntegration, OpenAI::Internal::Util::AnyHash), + wandb: T.any(OpenAI::Models::FineTuning::FineTuningJobWandbIntegration, OpenAI::Internal::AnyHash), type: Symbol ) .returns(T.attached_class) diff --git a/rbi/lib/openai/models/fine_tuning/job_cancel_params.rbi b/rbi/lib/openai/models/fine_tuning/job_cancel_params.rbi index 4180a266..2c2b8d46 100644 --- a/rbi/lib/openai/models/fine_tuning/job_cancel_params.rbi +++ b/rbi/lib/openai/models/fine_tuning/job_cancel_params.rbi @@ -3,13 +3,17 @@ module OpenAI module Models module FineTuning - class JobCancelParams < OpenAI::BaseModel + class JobCancelParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters sig do - params(request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) - .returns(T.attached_class) + params( + request_options: T.any( + OpenAI::RequestOptions, + OpenAI::Internal::AnyHash + ) + ).returns(T.attached_class) end def self.new(request_options: {}) end diff --git a/rbi/lib/openai/models/fine_tuning/job_create_params.rbi b/rbi/lib/openai/models/fine_tuning/job_create_params.rbi index 68089ea2..351f2408 100644 --- a/rbi/lib/openai/models/fine_tuning/job_create_params.rbi +++ b/rbi/lib/openai/models/fine_tuning/job_create_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module FineTuning - class JobCreateParams < OpenAI::BaseModel + class JobCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -39,7 +39,7 @@ module OpenAI sig do params( - hyperparameters: T.any(OpenAI::Models::FineTuning::JobCreateParams::Hyperparameters, OpenAI::Internal::Util::AnyHash) + hyperparameters: T.any(OpenAI::Models::FineTuning::JobCreateParams::Hyperparameters, OpenAI::Internal::AnyHash) ) .void end @@ -62,12 +62,7 @@ module OpenAI sig { returns(T.nilable(OpenAI::Models::FineTuning::JobCreateParams::Method)) } attr_reader :method_ - sig do - params( - method_: T.any(OpenAI::Models::FineTuning::JobCreateParams::Method, OpenAI::Internal::Util::AnyHash) - ) - .void - end + sig { params(method_: T.any(OpenAI::Models::FineTuning::JobCreateParams::Method, OpenAI::Internal::AnyHash)).void } attr_writer :method_ # The seed controls the reproducibility of the job. Passing in the same seed and @@ -103,16 +98,16 @@ module OpenAI params( model: T.any(String, OpenAI::Models::FineTuning::JobCreateParams::Model::OrSymbol), training_file: String, - hyperparameters: T.any(OpenAI::Models::FineTuning::JobCreateParams::Hyperparameters, OpenAI::Internal::Util::AnyHash), + hyperparameters: T.any(OpenAI::Models::FineTuning::JobCreateParams::Hyperparameters, OpenAI::Internal::AnyHash), integrations: T.nilable( - T::Array[T.any(OpenAI::Models::FineTuning::JobCreateParams::Integration, OpenAI::Internal::Util::AnyHash)] + T::Array[T.any(OpenAI::Models::FineTuning::JobCreateParams::Integration, OpenAI::Internal::AnyHash)] ), metadata: T.nilable(T::Hash[Symbol, String]), - method_: T.any(OpenAI::Models::FineTuning::JobCreateParams::Method, OpenAI::Internal::Util::AnyHash), + method_: T.any(OpenAI::Models::FineTuning::JobCreateParams::Method, OpenAI::Internal::AnyHash), seed: T.nilable(Integer), suffix: T.nilable(String), validation_file: T.nilable(String), - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -153,7 +148,7 @@ module OpenAI # The name of the model to fine-tune. You can select one of the # [supported models](https://platform.openai.com/docs/guides/fine-tuning#which-models-can-be-fine-tuned). module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, OpenAI::Models::FineTuning::JobCreateParams::Model::OrSymbol]) } def self.variants @@ -169,7 +164,7 @@ module OpenAI GPT_4O_MINI = T.let(:"gpt-4o-mini", OpenAI::Models::FineTuning::JobCreateParams::Model::TaggedSymbol) end - class Hyperparameters < OpenAI::BaseModel + class Hyperparameters < OpenAI::Internal::Type::BaseModel # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. sig { returns(T.nilable(T.any(Symbol, Integer))) } @@ -223,7 +218,7 @@ module OpenAI # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. module BatchSize - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([Symbol, Integer]) } def self.variants @@ -233,7 +228,7 @@ module OpenAI # Scaling factor for the learning rate. A smaller learning rate may be useful to # avoid overfitting. module LearningRateMultiplier - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([Symbol, Float]) } def self.variants @@ -243,7 +238,7 @@ module OpenAI # The number of epochs to train the model for. An epoch refers to one full cycle # through the training dataset. module NEpochs - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([Symbol, Integer]) } def self.variants @@ -251,7 +246,7 @@ module OpenAI end end - class Integration < OpenAI::BaseModel + class Integration < OpenAI::Internal::Type::BaseModel # The type of integration to enable. Currently, only "wandb" (Weights and Biases) # is supported. sig { returns(Symbol) } @@ -266,7 +261,7 @@ module OpenAI sig do params( - wandb: T.any(OpenAI::Models::FineTuning::JobCreateParams::Integration::Wandb, OpenAI::Internal::Util::AnyHash) + wandb: T.any(OpenAI::Models::FineTuning::JobCreateParams::Integration::Wandb, OpenAI::Internal::AnyHash) ) .void end @@ -274,7 +269,7 @@ module OpenAI sig do params( - wandb: T.any(OpenAI::Models::FineTuning::JobCreateParams::Integration::Wandb, OpenAI::Internal::Util::AnyHash), + wandb: T.any(OpenAI::Models::FineTuning::JobCreateParams::Integration::Wandb, OpenAI::Internal::AnyHash), type: Symbol ) .returns(T.attached_class) @@ -286,7 +281,7 @@ module OpenAI def to_hash end - class Wandb < OpenAI::BaseModel + class Wandb < OpenAI::Internal::Type::BaseModel # The name of the project that the new run will be created under. sig { returns(String) } attr_accessor :project @@ -341,15 +336,13 @@ module OpenAI end end - class Method < OpenAI::BaseModel + class Method < OpenAI::Internal::Type::BaseModel # Configuration for the DPO fine-tuning method. sig { returns(T.nilable(OpenAI::Models::FineTuning::JobCreateParams::Method::Dpo)) } attr_reader :dpo sig do - params( - dpo: T.any(OpenAI::Models::FineTuning::JobCreateParams::Method::Dpo, OpenAI::Internal::Util::AnyHash) - ) + params(dpo: T.any(OpenAI::Models::FineTuning::JobCreateParams::Method::Dpo, OpenAI::Internal::AnyHash)) .void end attr_writer :dpo @@ -360,7 +353,7 @@ module OpenAI sig do params( - supervised: T.any(OpenAI::Models::FineTuning::JobCreateParams::Method::Supervised, OpenAI::Internal::Util::AnyHash) + supervised: T.any(OpenAI::Models::FineTuning::JobCreateParams::Method::Supervised, OpenAI::Internal::AnyHash) ) .void end @@ -376,8 +369,8 @@ module OpenAI # The method used for fine-tuning. sig do params( - dpo: T.any(OpenAI::Models::FineTuning::JobCreateParams::Method::Dpo, OpenAI::Internal::Util::AnyHash), - supervised: T.any(OpenAI::Models::FineTuning::JobCreateParams::Method::Supervised, OpenAI::Internal::Util::AnyHash), + dpo: T.any(OpenAI::Models::FineTuning::JobCreateParams::Method::Dpo, OpenAI::Internal::AnyHash), + supervised: T.any(OpenAI::Models::FineTuning::JobCreateParams::Method::Supervised, OpenAI::Internal::AnyHash), type: OpenAI::Models::FineTuning::JobCreateParams::Method::Type::OrSymbol ) .returns(T.attached_class) @@ -398,7 +391,7 @@ module OpenAI def to_hash end - class Dpo < OpenAI::BaseModel + class Dpo < OpenAI::Internal::Type::BaseModel # The hyperparameters used for the fine-tuning job. sig { returns(T.nilable(OpenAI::Models::FineTuning::JobCreateParams::Method::Dpo::Hyperparameters)) } attr_reader :hyperparameters @@ -407,7 +400,7 @@ module OpenAI params( hyperparameters: T.any( OpenAI::Models::FineTuning::JobCreateParams::Method::Dpo::Hyperparameters, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .void @@ -419,7 +412,7 @@ module OpenAI params( hyperparameters: T.any( OpenAI::Models::FineTuning::JobCreateParams::Method::Dpo::Hyperparameters, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .returns(T.attached_class) @@ -434,7 +427,7 @@ module OpenAI def to_hash end - class Hyperparameters < OpenAI::BaseModel + class Hyperparameters < OpenAI::Internal::Type::BaseModel # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. sig { returns(T.nilable(T.any(Symbol, Integer))) } @@ -497,7 +490,7 @@ module OpenAI # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. module BatchSize - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([Symbol, Integer]) } def self.variants @@ -507,7 +500,7 @@ module OpenAI # The beta value for the DPO method. A higher beta value will increase the weight # of the penalty between the policy and reference model. module Beta - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([Symbol, Float]) } def self.variants @@ -517,7 +510,7 @@ module OpenAI # Scaling factor for the learning rate. A smaller learning rate may be useful to # avoid overfitting. module LearningRateMultiplier - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([Symbol, Float]) } def self.variants @@ -527,7 +520,7 @@ module OpenAI # The number of epochs to train the model for. An epoch refers to one full cycle # through the training dataset. module NEpochs - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([Symbol, Integer]) } def self.variants @@ -536,7 +529,7 @@ module OpenAI end end - class Supervised < OpenAI::BaseModel + class Supervised < OpenAI::Internal::Type::BaseModel # The hyperparameters used for the fine-tuning job. sig { returns(T.nilable(OpenAI::Models::FineTuning::JobCreateParams::Method::Supervised::Hyperparameters)) } attr_reader :hyperparameters @@ -545,7 +538,7 @@ module OpenAI params( hyperparameters: T.any( OpenAI::Models::FineTuning::JobCreateParams::Method::Supervised::Hyperparameters, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .void @@ -557,7 +550,7 @@ module OpenAI params( hyperparameters: T.any( OpenAI::Models::FineTuning::JobCreateParams::Method::Supervised::Hyperparameters, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ) .returns(T.attached_class) @@ -574,7 +567,7 @@ module OpenAI def to_hash end - class Hyperparameters < OpenAI::BaseModel + class Hyperparameters < OpenAI::Internal::Type::BaseModel # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. sig { returns(T.nilable(T.any(Symbol, Integer))) } @@ -627,7 +620,7 @@ module OpenAI # Number of examples in each batch. A larger batch size means that model # parameters are updated less frequently, but with lower variance. module BatchSize - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([Symbol, Integer]) } def self.variants @@ -637,7 +630,7 @@ module OpenAI # Scaling factor for the learning rate. A smaller learning rate may be useful to # avoid overfitting. module LearningRateMultiplier - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([Symbol, Float]) } def self.variants @@ -647,7 +640,7 @@ module OpenAI # The number of epochs to train the model for. An epoch refers to one full cycle # through the training dataset. module NEpochs - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([Symbol, Integer]) } def self.variants @@ -658,7 +651,7 @@ module OpenAI # The type of method. Is either `supervised` or `dpo`. module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::FineTuning::JobCreateParams::Method::Type) } OrSymbol = diff --git a/rbi/lib/openai/models/fine_tuning/job_list_events_params.rbi b/rbi/lib/openai/models/fine_tuning/job_list_events_params.rbi index 5a289277..9ad33668 100644 --- a/rbi/lib/openai/models/fine_tuning/job_list_events_params.rbi +++ b/rbi/lib/openai/models/fine_tuning/job_list_events_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module FineTuning - class JobListEventsParams < OpenAI::BaseModel + class JobListEventsParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -25,7 +25,7 @@ module OpenAI params( after: String, limit: Integer, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end diff --git a/rbi/lib/openai/models/fine_tuning/job_list_params.rbi b/rbi/lib/openai/models/fine_tuning/job_list_params.rbi index d3d055fc..94aa1f34 100644 --- a/rbi/lib/openai/models/fine_tuning/job_list_params.rbi +++ b/rbi/lib/openai/models/fine_tuning/job_list_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module FineTuning - class JobListParams < OpenAI::BaseModel + class JobListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -31,7 +31,7 @@ module OpenAI after: String, limit: Integer, metadata: T.nilable(T::Hash[Symbol, String]), - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end diff --git a/rbi/lib/openai/models/fine_tuning/job_retrieve_params.rbi b/rbi/lib/openai/models/fine_tuning/job_retrieve_params.rbi index 119b93f5..ae440b7a 100644 --- a/rbi/lib/openai/models/fine_tuning/job_retrieve_params.rbi +++ b/rbi/lib/openai/models/fine_tuning/job_retrieve_params.rbi @@ -3,13 +3,17 @@ module OpenAI module Models module FineTuning - class JobRetrieveParams < OpenAI::BaseModel + class JobRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters sig do - params(request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) - .returns(T.attached_class) + params( + request_options: T.any( + OpenAI::RequestOptions, + OpenAI::Internal::AnyHash + ) + ).returns(T.attached_class) end def self.new(request_options: {}) end diff --git a/rbi/lib/openai/models/fine_tuning/jobs/checkpoint_list_params.rbi b/rbi/lib/openai/models/fine_tuning/jobs/checkpoint_list_params.rbi index e9dabed3..279608d3 100644 --- a/rbi/lib/openai/models/fine_tuning/jobs/checkpoint_list_params.rbi +++ b/rbi/lib/openai/models/fine_tuning/jobs/checkpoint_list_params.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module FineTuning module Jobs - class CheckpointListParams < OpenAI::BaseModel + class CheckpointListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -26,7 +26,7 @@ module OpenAI params( after: String, limit: Integer, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end diff --git a/rbi/lib/openai/models/fine_tuning/jobs/fine_tuning_job_checkpoint.rbi b/rbi/lib/openai/models/fine_tuning/jobs/fine_tuning_job_checkpoint.rbi index f51faa86..cd3931c5 100644 --- a/rbi/lib/openai/models/fine_tuning/jobs/fine_tuning_job_checkpoint.rbi +++ b/rbi/lib/openai/models/fine_tuning/jobs/fine_tuning_job_checkpoint.rbi @@ -4,7 +4,7 @@ module OpenAI module Models module FineTuning module Jobs - class FineTuningJobCheckpoint < OpenAI::BaseModel + class FineTuningJobCheckpoint < OpenAI::Internal::Type::BaseModel # The checkpoint identifier, which can be referenced in the API endpoints. sig { returns(String) } attr_accessor :id @@ -27,7 +27,7 @@ module OpenAI sig do params( - metrics: T.any(OpenAI::Models::FineTuning::Jobs::FineTuningJobCheckpoint::Metrics, OpenAI::Internal::Util::AnyHash) + metrics: T.any(OpenAI::Models::FineTuning::Jobs::FineTuningJobCheckpoint::Metrics, OpenAI::Internal::AnyHash) ) .void end @@ -49,7 +49,7 @@ module OpenAI created_at: Integer, fine_tuned_model_checkpoint: String, fine_tuning_job_id: String, - metrics: T.any(OpenAI::Models::FineTuning::Jobs::FineTuningJobCheckpoint::Metrics, OpenAI::Internal::Util::AnyHash), + metrics: T.any(OpenAI::Models::FineTuning::Jobs::FineTuningJobCheckpoint::Metrics, OpenAI::Internal::AnyHash), step_number: Integer, object: Symbol ) @@ -83,7 +83,7 @@ module OpenAI def to_hash end - class Metrics < OpenAI::BaseModel + class Metrics < OpenAI::Internal::Type::BaseModel sig { returns(T.nilable(Float)) } attr_reader :full_valid_loss diff --git a/rbi/lib/openai/models/function_definition.rbi b/rbi/lib/openai/models/function_definition.rbi index f1bad3a8..6a4dc07a 100644 --- a/rbi/lib/openai/models/function_definition.rbi +++ b/rbi/lib/openai/models/function_definition.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class FunctionDefinition < OpenAI::BaseModel + class FunctionDefinition < OpenAI::Internal::Type::BaseModel # The name of the function to be called. Must be a-z, A-Z, 0-9, or contain # underscores and dashes, with a maximum length of 64. sig { returns(String) } diff --git a/rbi/lib/openai/models/function_parameters.rbi b/rbi/lib/openai/models/function_parameters.rbi index 5920964f..f7b78f39 100644 --- a/rbi/lib/openai/models/function_parameters.rbi +++ b/rbi/lib/openai/models/function_parameters.rbi @@ -2,6 +2,10 @@ module OpenAI module Models - FunctionParameters = T.let(OpenAI::HashOf[OpenAI::Unknown], OpenAI::Internal::Type::Converter) + FunctionParameters = + T.let( + OpenAI::Internal::Type::HashOf[OpenAI::Internal::Type::Unknown], + OpenAI::Internal::Type::Converter + ) end end diff --git a/rbi/lib/openai/models/image.rbi b/rbi/lib/openai/models/image.rbi index 5af93081..9f14f13d 100644 --- a/rbi/lib/openai/models/image.rbi +++ b/rbi/lib/openai/models/image.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class Image < OpenAI::BaseModel + class Image < OpenAI::Internal::Type::BaseModel # The base64-encoded JSON of the generated image, if `response_format` is # `b64_json`. sig { returns(T.nilable(String)) } diff --git a/rbi/lib/openai/models/image_create_variation_params.rbi b/rbi/lib/openai/models/image_create_variation_params.rbi index ea1298ad..eb1c580e 100644 --- a/rbi/lib/openai/models/image_create_variation_params.rbi +++ b/rbi/lib/openai/models/image_create_variation_params.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class ImageCreateVariationParams < OpenAI::BaseModel + class ImageCreateVariationParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -49,7 +49,7 @@ module OpenAI response_format: T.nilable(OpenAI::Models::ImageCreateVariationParams::ResponseFormat::OrSymbol), size: T.nilable(OpenAI::Models::ImageCreateVariationParams::Size::OrSymbol), user: String, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -84,7 +84,7 @@ module OpenAI # The model to use for image generation. Only `dall-e-2` is supported at this # time. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, OpenAI::Models::ImageModel::OrSymbol]) } def self.variants @@ -95,7 +95,7 @@ module OpenAI # `b64_json`. URLs are only valid for 60 minutes after the image has been # generated. module ResponseFormat - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::ImageCreateVariationParams::ResponseFormat) } OrSymbol = @@ -112,7 +112,7 @@ module OpenAI # The size of the generated images. Must be one of `256x256`, `512x512`, or # `1024x1024`. module Size - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::ImageCreateVariationParams::Size) } OrSymbol = diff --git a/rbi/lib/openai/models/image_edit_params.rbi b/rbi/lib/openai/models/image_edit_params.rbi index 80ac6fdf..5c428b9a 100644 --- a/rbi/lib/openai/models/image_edit_params.rbi +++ b/rbi/lib/openai/models/image_edit_params.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class ImageEditParams < OpenAI::BaseModel + class ImageEditParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -64,7 +64,7 @@ module OpenAI response_format: T.nilable(OpenAI::Models::ImageEditParams::ResponseFormat::OrSymbol), size: T.nilable(OpenAI::Models::ImageEditParams::Size::OrSymbol), user: String, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -103,7 +103,7 @@ module OpenAI # The model to use for image generation. Only `dall-e-2` is supported at this # time. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, OpenAI::Models::ImageModel::OrSymbol]) } def self.variants @@ -114,7 +114,7 @@ module OpenAI # `b64_json`. URLs are only valid for 60 minutes after the image has been # generated. module ResponseFormat - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::ImageEditParams::ResponseFormat) } OrSymbol = @@ -131,7 +131,7 @@ module OpenAI # The size of the generated images. Must be one of `256x256`, `512x512`, or # `1024x1024`. module Size - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::ImageEditParams::Size) } OrSymbol = T.type_alias { T.any(Symbol, String, OpenAI::Models::ImageEditParams::Size::TaggedSymbol) } diff --git a/rbi/lib/openai/models/image_generate_params.rbi b/rbi/lib/openai/models/image_generate_params.rbi index 5b0720c6..880ba871 100644 --- a/rbi/lib/openai/models/image_generate_params.rbi +++ b/rbi/lib/openai/models/image_generate_params.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class ImageGenerateParams < OpenAI::BaseModel + class ImageGenerateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -67,7 +67,7 @@ module OpenAI size: T.nilable(OpenAI::Models::ImageGenerateParams::Size::OrSymbol), style: T.nilable(OpenAI::Models::ImageGenerateParams::Style::OrSymbol), user: String, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -105,7 +105,7 @@ module OpenAI # The model to use for image generation. module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, OpenAI::Models::ImageModel::OrSymbol]) } def self.variants @@ -116,7 +116,7 @@ module OpenAI # details and greater consistency across the image. This param is only supported # for `dall-e-3`. module Quality - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::ImageGenerateParams::Quality) } OrSymbol = @@ -134,7 +134,7 @@ module OpenAI # `b64_json`. URLs are only valid for 60 minutes after the image has been # generated. module ResponseFormat - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::ImageGenerateParams::ResponseFormat) } OrSymbol = @@ -152,7 +152,7 @@ module OpenAI # `1024x1024` for `dall-e-2`. Must be one of `1024x1024`, `1792x1024`, or # `1024x1792` for `dall-e-3` models. module Size - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::ImageGenerateParams::Size) } OrSymbol = @@ -174,7 +174,7 @@ module OpenAI # Natural causes the model to produce more natural, less hyper-real looking # images. This param is only supported for `dall-e-3`. module Style - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::ImageGenerateParams::Style) } OrSymbol = diff --git a/rbi/lib/openai/models/image_model.rbi b/rbi/lib/openai/models/image_model.rbi index 6ba143ea..20027c2b 100644 --- a/rbi/lib/openai/models/image_model.rbi +++ b/rbi/lib/openai/models/image_model.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module ImageModel - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::ImageModel) } OrSymbol = T.type_alias { T.any(Symbol, String, OpenAI::Models::ImageModel::TaggedSymbol) } diff --git a/rbi/lib/openai/models/images_response.rbi b/rbi/lib/openai/models/images_response.rbi index 412cb40b..796ce49c 100644 --- a/rbi/lib/openai/models/images_response.rbi +++ b/rbi/lib/openai/models/images_response.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class ImagesResponse < OpenAI::BaseModel + class ImagesResponse < OpenAI::Internal::Type::BaseModel sig { returns(Integer) } attr_accessor :created @@ -10,7 +10,7 @@ module OpenAI attr_accessor :data sig do - params(created: Integer, data: T::Array[T.any(OpenAI::Models::Image, OpenAI::Internal::Util::AnyHash)]) + params(created: Integer, data: T::Array[T.any(OpenAI::Models::Image, OpenAI::Internal::AnyHash)]) .returns(T.attached_class) end def self.new(created:, data:) diff --git a/rbi/lib/openai/models/metadata.rbi b/rbi/lib/openai/models/metadata.rbi index 20e8347e..04102855 100644 --- a/rbi/lib/openai/models/metadata.rbi +++ b/rbi/lib/openai/models/metadata.rbi @@ -2,6 +2,6 @@ module OpenAI module Models - Metadata = T.let(OpenAI::HashOf[String], OpenAI::Internal::Type::Converter) + Metadata = T.let(OpenAI::Internal::Type::HashOf[String], OpenAI::Internal::Type::Converter) end end diff --git a/rbi/lib/openai/models/model.rbi b/rbi/lib/openai/models/model.rbi index 556bb5fc..ad55cfee 100644 --- a/rbi/lib/openai/models/model.rbi +++ b/rbi/lib/openai/models/model.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class Model < OpenAI::BaseModel + class Model < OpenAI::Internal::Type::BaseModel # The model identifier, which can be referenced in the API endpoints. sig { returns(String) } attr_accessor :id diff --git a/rbi/lib/openai/models/model_delete_params.rbi b/rbi/lib/openai/models/model_delete_params.rbi index 3ace3c62..42095a36 100644 --- a/rbi/lib/openai/models/model_delete_params.rbi +++ b/rbi/lib/openai/models/model_delete_params.rbi @@ -2,13 +2,17 @@ module OpenAI module Models - class ModelDeleteParams < OpenAI::BaseModel + class ModelDeleteParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters sig do - params(request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) - .returns(T.attached_class) + params( + request_options: T.any( + OpenAI::RequestOptions, + OpenAI::Internal::AnyHash + ) + ).returns(T.attached_class) end def self.new(request_options: {}) end diff --git a/rbi/lib/openai/models/model_deleted.rbi b/rbi/lib/openai/models/model_deleted.rbi index 9888bb84..7044add0 100644 --- a/rbi/lib/openai/models/model_deleted.rbi +++ b/rbi/lib/openai/models/model_deleted.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class ModelDeleted < OpenAI::BaseModel + class ModelDeleted < OpenAI::Internal::Type::BaseModel sig { returns(String) } attr_accessor :id diff --git a/rbi/lib/openai/models/model_list_params.rbi b/rbi/lib/openai/models/model_list_params.rbi index ec503d0e..6570ac6f 100644 --- a/rbi/lib/openai/models/model_list_params.rbi +++ b/rbi/lib/openai/models/model_list_params.rbi @@ -2,13 +2,17 @@ module OpenAI module Models - class ModelListParams < OpenAI::BaseModel + class ModelListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters sig do - params(request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) - .returns(T.attached_class) + params( + request_options: T.any( + OpenAI::RequestOptions, + OpenAI::Internal::AnyHash + ) + ).returns(T.attached_class) end def self.new(request_options: {}) end diff --git a/rbi/lib/openai/models/model_retrieve_params.rbi b/rbi/lib/openai/models/model_retrieve_params.rbi index 48ab1cf7..273f1b0e 100644 --- a/rbi/lib/openai/models/model_retrieve_params.rbi +++ b/rbi/lib/openai/models/model_retrieve_params.rbi @@ -2,13 +2,17 @@ module OpenAI module Models - class ModelRetrieveParams < OpenAI::BaseModel + class ModelRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters sig do - params(request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) - .returns(T.attached_class) + params( + request_options: T.any( + OpenAI::RequestOptions, + OpenAI::Internal::AnyHash + ) + ).returns(T.attached_class) end def self.new(request_options: {}) end diff --git a/rbi/lib/openai/models/moderation.rbi b/rbi/lib/openai/models/moderation.rbi index 256ad93b..70df9dfa 100644 --- a/rbi/lib/openai/models/moderation.rbi +++ b/rbi/lib/openai/models/moderation.rbi @@ -2,12 +2,12 @@ module OpenAI module Models - class Moderation < OpenAI::BaseModel + class Moderation < OpenAI::Internal::Type::BaseModel # A list of the categories, and whether they are flagged or not. sig { returns(OpenAI::Models::Moderation::Categories) } attr_reader :categories - sig { params(categories: T.any(OpenAI::Models::Moderation::Categories, OpenAI::Internal::Util::AnyHash)).void } + sig { params(categories: T.any(OpenAI::Models::Moderation::Categories, OpenAI::Internal::AnyHash)).void } attr_writer :categories # A list of the categories along with the input type(s) that the score applies to. @@ -16,7 +16,7 @@ module OpenAI sig do params( - category_applied_input_types: T.any(OpenAI::Models::Moderation::CategoryAppliedInputTypes, OpenAI::Internal::Util::AnyHash) + category_applied_input_types: T.any(OpenAI::Models::Moderation::CategoryAppliedInputTypes, OpenAI::Internal::AnyHash) ) .void end @@ -26,12 +26,7 @@ module OpenAI sig { returns(OpenAI::Models::Moderation::CategoryScores) } attr_reader :category_scores - sig do - params( - category_scores: T.any(OpenAI::Models::Moderation::CategoryScores, OpenAI::Internal::Util::AnyHash) - ) - .void - end + sig { params(category_scores: T.any(OpenAI::Models::Moderation::CategoryScores, OpenAI::Internal::AnyHash)).void } attr_writer :category_scores # Whether any of the below categories are flagged. @@ -40,9 +35,9 @@ module OpenAI sig do params( - categories: T.any(OpenAI::Models::Moderation::Categories, OpenAI::Internal::Util::AnyHash), - category_applied_input_types: T.any(OpenAI::Models::Moderation::CategoryAppliedInputTypes, OpenAI::Internal::Util::AnyHash), - category_scores: T.any(OpenAI::Models::Moderation::CategoryScores, OpenAI::Internal::Util::AnyHash), + categories: T.any(OpenAI::Models::Moderation::Categories, OpenAI::Internal::AnyHash), + category_applied_input_types: T.any(OpenAI::Models::Moderation::CategoryAppliedInputTypes, OpenAI::Internal::AnyHash), + category_scores: T.any(OpenAI::Models::Moderation::CategoryScores, OpenAI::Internal::AnyHash), flagged: T::Boolean ) .returns(T.attached_class) @@ -64,7 +59,7 @@ module OpenAI def to_hash end - class Categories < OpenAI::BaseModel + class Categories < OpenAI::Internal::Type::BaseModel # Content that expresses, incites, or promotes harassing language towards any # target. sig { returns(T::Boolean) } @@ -194,7 +189,7 @@ module OpenAI end end - class CategoryAppliedInputTypes < OpenAI::BaseModel + class CategoryAppliedInputTypes < OpenAI::Internal::Type::BaseModel # The applied input type(s) for the category 'harassment'. sig { returns(T::Array[OpenAI::Models::Moderation::CategoryAppliedInputTypes::Harassment::TaggedSymbol]) } attr_accessor :harassment @@ -315,7 +310,7 @@ module OpenAI end module Harassment - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Moderation::CategoryAppliedInputTypes::Harassment) } @@ -332,7 +327,7 @@ module OpenAI end module HarassmentThreatening - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Moderation::CategoryAppliedInputTypes::HarassmentThreatening) } @@ -359,7 +354,7 @@ module OpenAI end module Hate - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Moderation::CategoryAppliedInputTypes::Hate) } @@ -374,7 +369,7 @@ module OpenAI end module HateThreatening - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Moderation::CategoryAppliedInputTypes::HateThreatening) } @@ -398,7 +393,7 @@ module OpenAI end module Illicit - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Moderation::CategoryAppliedInputTypes::Illicit) } @@ -413,7 +408,7 @@ module OpenAI end module IllicitViolent - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Moderation::CategoryAppliedInputTypes::IllicitViolent) } @@ -431,7 +426,7 @@ module OpenAI end module SelfHarm - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Moderation::CategoryAppliedInputTypes::SelfHarm) } @@ -447,7 +442,7 @@ module OpenAI end module SelfHarmInstruction - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Moderation::CategoryAppliedInputTypes::SelfHarmInstruction) } @@ -476,7 +471,7 @@ module OpenAI end module SelfHarmIntent - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Moderation::CategoryAppliedInputTypes::SelfHarmIntent) } @@ -496,7 +491,7 @@ module OpenAI end module Sexual - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Moderation::CategoryAppliedInputTypes::Sexual) } @@ -512,7 +507,7 @@ module OpenAI end module SexualMinor - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Moderation::CategoryAppliedInputTypes::SexualMinor) } @@ -530,7 +525,7 @@ module OpenAI end module Violence - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Moderation::CategoryAppliedInputTypes::Violence) } @@ -546,7 +541,7 @@ module OpenAI end module ViolenceGraphic - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Moderation::CategoryAppliedInputTypes::ViolenceGraphic) } @@ -572,7 +567,7 @@ module OpenAI end end - class CategoryScores < OpenAI::BaseModel + class CategoryScores < OpenAI::Internal::Type::BaseModel # The score for the category 'harassment'. sig { returns(Float) } attr_accessor :harassment diff --git a/rbi/lib/openai/models/moderation_create_params.rbi b/rbi/lib/openai/models/moderation_create_params.rbi index c66af1c0..4fa9650a 100644 --- a/rbi/lib/openai/models/moderation_create_params.rbi +++ b/rbi/lib/openai/models/moderation_create_params.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class ModerationCreateParams < OpenAI::BaseModel + class ModerationCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -37,13 +37,13 @@ module OpenAI T::Array[ T.any( OpenAI::Models::ModerationImageURLInput, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::ModerationTextInput ) ] ), model: T.any(String, OpenAI::Models::ModerationModel::OrSymbol), - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -70,7 +70,7 @@ module OpenAI # Input (or inputs) to classify. Can be a single string, an array of strings, or # an array of multi-modal input objects similar to other models. module Input - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override @@ -81,11 +81,11 @@ module OpenAI def self.variants end - StringArray = T.let(OpenAI::ArrayOf[String], OpenAI::Internal::Type::Converter) + StringArray = T.let(OpenAI::Internal::Type::ArrayOf[String], OpenAI::Internal::Type::Converter) ModerationMultiModalInputArray = T.let( - OpenAI::ArrayOf[union: OpenAI::Models::ModerationMultiModalInput], + OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::ModerationMultiModalInput], OpenAI::Internal::Type::Converter ) end @@ -95,7 +95,7 @@ module OpenAI # learn about available models # [here](https://platform.openai.com/docs/models#moderation). module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, OpenAI::Models::ModerationModel::OrSymbol]) } def self.variants diff --git a/rbi/lib/openai/models/moderation_create_response.rbi b/rbi/lib/openai/models/moderation_create_response.rbi index 4c1b0187..fc12e123 100644 --- a/rbi/lib/openai/models/moderation_create_response.rbi +++ b/rbi/lib/openai/models/moderation_create_response.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class ModerationCreateResponse < OpenAI::BaseModel + class ModerationCreateResponse < OpenAI::Internal::Type::BaseModel # The unique identifier for the moderation request. sig { returns(String) } attr_accessor :id @@ -20,7 +20,7 @@ module OpenAI params( id: String, model: String, - results: T::Array[T.any(OpenAI::Models::Moderation, OpenAI::Internal::Util::AnyHash)] + results: T::Array[T.any(OpenAI::Models::Moderation, OpenAI::Internal::AnyHash)] ) .returns(T.attached_class) end diff --git a/rbi/lib/openai/models/moderation_image_url_input.rbi b/rbi/lib/openai/models/moderation_image_url_input.rbi index 77f4cfc0..128f5062 100644 --- a/rbi/lib/openai/models/moderation_image_url_input.rbi +++ b/rbi/lib/openai/models/moderation_image_url_input.rbi @@ -2,17 +2,12 @@ module OpenAI module Models - class ModerationImageURLInput < OpenAI::BaseModel + class ModerationImageURLInput < OpenAI::Internal::Type::BaseModel # Contains either an image URL or a data URL for a base64 encoded image. sig { returns(OpenAI::Models::ModerationImageURLInput::ImageURL) } attr_reader :image_url - sig do - params( - image_url: T.any(OpenAI::Models::ModerationImageURLInput::ImageURL, OpenAI::Internal::Util::AnyHash) - ) - .void - end + sig { params(image_url: T.any(OpenAI::Models::ModerationImageURLInput::ImageURL, OpenAI::Internal::AnyHash)).void } attr_writer :image_url # Always `image_url`. @@ -22,7 +17,7 @@ module OpenAI # An object describing an image to classify. sig do params( - image_url: T.any(OpenAI::Models::ModerationImageURLInput::ImageURL, OpenAI::Internal::Util::AnyHash), + image_url: T.any(OpenAI::Models::ModerationImageURLInput::ImageURL, OpenAI::Internal::AnyHash), type: Symbol ) .returns(T.attached_class) @@ -34,7 +29,7 @@ module OpenAI def to_hash end - class ImageURL < OpenAI::BaseModel + class ImageURL < OpenAI::Internal::Type::BaseModel # Either a URL of the image or the base64 encoded image data. sig { returns(String) } attr_accessor :url diff --git a/rbi/lib/openai/models/moderation_model.rbi b/rbi/lib/openai/models/moderation_model.rbi index e422afd9..3b1d564a 100644 --- a/rbi/lib/openai/models/moderation_model.rbi +++ b/rbi/lib/openai/models/moderation_model.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module ModerationModel - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::ModerationModel) } OrSymbol = T.type_alias { T.any(Symbol, String, OpenAI::Models::ModerationModel::TaggedSymbol) } diff --git a/rbi/lib/openai/models/moderation_multi_modal_input.rbi b/rbi/lib/openai/models/moderation_multi_modal_input.rbi index 9c96d7b0..247b0857 100644 --- a/rbi/lib/openai/models/moderation_multi_modal_input.rbi +++ b/rbi/lib/openai/models/moderation_multi_modal_input.rbi @@ -4,7 +4,7 @@ module OpenAI module Models # An object describing an image to classify. module ModerationMultiModalInput - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([OpenAI::Models::ModerationImageURLInput, OpenAI::Models::ModerationTextInput]) } def self.variants diff --git a/rbi/lib/openai/models/moderation_text_input.rbi b/rbi/lib/openai/models/moderation_text_input.rbi index dddf6d0b..d117af19 100644 --- a/rbi/lib/openai/models/moderation_text_input.rbi +++ b/rbi/lib/openai/models/moderation_text_input.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class ModerationTextInput < OpenAI::BaseModel + class ModerationTextInput < OpenAI::Internal::Type::BaseModel # A string of text to classify. sig { returns(String) } attr_accessor :text diff --git a/rbi/lib/openai/models/other_file_chunking_strategy_object.rbi b/rbi/lib/openai/models/other_file_chunking_strategy_object.rbi index 14da388e..24704550 100644 --- a/rbi/lib/openai/models/other_file_chunking_strategy_object.rbi +++ b/rbi/lib/openai/models/other_file_chunking_strategy_object.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class OtherFileChunkingStrategyObject < OpenAI::BaseModel + class OtherFileChunkingStrategyObject < OpenAI::Internal::Type::BaseModel # Always `other`. sig { returns(Symbol) } attr_accessor :type diff --git a/rbi/lib/openai/models/reasoning.rbi b/rbi/lib/openai/models/reasoning.rbi index 0bcde269..13f6e5ae 100644 --- a/rbi/lib/openai/models/reasoning.rbi +++ b/rbi/lib/openai/models/reasoning.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class Reasoning < OpenAI::BaseModel + class Reasoning < OpenAI::Internal::Type::BaseModel # **o-series models only** # # Constrains effort on reasoning for @@ -52,7 +52,7 @@ module OpenAI # debugging and understanding the model's reasoning process. One of `concise` or # `detailed`. module GenerateSummary - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Reasoning::GenerateSummary) } OrSymbol = diff --git a/rbi/lib/openai/models/reasoning_effort.rbi b/rbi/lib/openai/models/reasoning_effort.rbi index e95a9cdc..d46cfdf9 100644 --- a/rbi/lib/openai/models/reasoning_effort.rbi +++ b/rbi/lib/openai/models/reasoning_effort.rbi @@ -9,7 +9,7 @@ module OpenAI # supported values are `low`, `medium`, and `high`. Reducing reasoning effort can # result in faster responses and fewer tokens used on reasoning in a response. module ReasoningEffort - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::ReasoningEffort) } OrSymbol = T.type_alias { T.any(Symbol, String, OpenAI::Models::ReasoningEffort::TaggedSymbol) } diff --git a/rbi/lib/openai/models/response_format_json_object.rbi b/rbi/lib/openai/models/response_format_json_object.rbi index 331cb19e..3a1e8d74 100644 --- a/rbi/lib/openai/models/response_format_json_object.rbi +++ b/rbi/lib/openai/models/response_format_json_object.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class ResponseFormatJSONObject < OpenAI::BaseModel + class ResponseFormatJSONObject < OpenAI::Internal::Type::BaseModel # The type of response format being defined. Always `json_object`. sig { returns(Symbol) } attr_accessor :type diff --git a/rbi/lib/openai/models/response_format_json_schema.rbi b/rbi/lib/openai/models/response_format_json_schema.rbi index 95e6df84..8f5dbd5c 100644 --- a/rbi/lib/openai/models/response_format_json_schema.rbi +++ b/rbi/lib/openai/models/response_format_json_schema.rbi @@ -2,14 +2,14 @@ module OpenAI module Models - class ResponseFormatJSONSchema < OpenAI::BaseModel + class ResponseFormatJSONSchema < OpenAI::Internal::Type::BaseModel # Structured Outputs configuration options, including a JSON Schema. sig { returns(OpenAI::Models::ResponseFormatJSONSchema::JSONSchema) } attr_reader :json_schema sig do params( - json_schema: T.any(OpenAI::Models::ResponseFormatJSONSchema::JSONSchema, OpenAI::Internal::Util::AnyHash) + json_schema: T.any(OpenAI::Models::ResponseFormatJSONSchema::JSONSchema, OpenAI::Internal::AnyHash) ) .void end @@ -24,7 +24,7 @@ module OpenAI # [Structured Outputs](https://platform.openai.com/docs/guides/structured-outputs). sig do params( - json_schema: T.any(OpenAI::Models::ResponseFormatJSONSchema::JSONSchema, OpenAI::Internal::Util::AnyHash), + json_schema: T.any(OpenAI::Models::ResponseFormatJSONSchema::JSONSchema, OpenAI::Internal::AnyHash), type: Symbol ) .returns(T.attached_class) @@ -36,7 +36,7 @@ module OpenAI def to_hash end - class JSONSchema < OpenAI::BaseModel + class JSONSchema < OpenAI::Internal::Type::BaseModel # The name of the response format. Must be a-z, A-Z, 0-9, or contain underscores # and dashes, with a maximum length of 64. sig { returns(String) } diff --git a/rbi/lib/openai/models/response_format_text.rbi b/rbi/lib/openai/models/response_format_text.rbi index a4b2db5e..2cee5b0b 100644 --- a/rbi/lib/openai/models/response_format_text.rbi +++ b/rbi/lib/openai/models/response_format_text.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class ResponseFormatText < OpenAI::BaseModel + class ResponseFormatText < OpenAI::Internal::Type::BaseModel # The type of response format being defined. Always `text`. sig { returns(Symbol) } attr_accessor :type diff --git a/rbi/lib/openai/models/responses/computer_tool.rbi b/rbi/lib/openai/models/responses/computer_tool.rbi index 90cabbf7..b4329d44 100644 --- a/rbi/lib/openai/models/responses/computer_tool.rbi +++ b/rbi/lib/openai/models/responses/computer_tool.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ComputerTool < OpenAI::BaseModel + class ComputerTool < OpenAI::Internal::Type::BaseModel # The height of the computer display. sig { returns(Float) } attr_accessor :display_height @@ -50,7 +50,7 @@ module OpenAI # The type of computer environment to control. module Environment - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ComputerTool::Environment) } OrSymbol = diff --git a/rbi/lib/openai/models/responses/easy_input_message.rbi b/rbi/lib/openai/models/responses/easy_input_message.rbi index 67ac236f..6b25a120 100644 --- a/rbi/lib/openai/models/responses/easy_input_message.rbi +++ b/rbi/lib/openai/models/responses/easy_input_message.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class EasyInputMessage < OpenAI::BaseModel + class EasyInputMessage < OpenAI::Internal::Type::BaseModel # Text, image, or audio input to the model, used to generate a response. Can also # contain previous assistant responses. sig do @@ -46,7 +46,7 @@ module OpenAI T::Array[ T.any( OpenAI::Models::Responses::ResponseInputText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ResponseInputImage, OpenAI::Models::Responses::ResponseInputFile ) @@ -85,7 +85,7 @@ module OpenAI # Text, image, or audio input to the model, used to generate a response. Can also # contain previous assistant responses. module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override @@ -109,7 +109,7 @@ module OpenAI # The role of the message input. One of `user`, `assistant`, `system`, or # `developer`. module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::EasyInputMessage::Role) } OrSymbol = @@ -127,7 +127,7 @@ module OpenAI # The type of the message input. Always `message`. module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::EasyInputMessage::Type) } OrSymbol = diff --git a/rbi/lib/openai/models/responses/file_search_tool.rbi b/rbi/lib/openai/models/responses/file_search_tool.rbi index 94dca043..7ca909fa 100644 --- a/rbi/lib/openai/models/responses/file_search_tool.rbi +++ b/rbi/lib/openai/models/responses/file_search_tool.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class FileSearchTool < OpenAI::BaseModel + class FileSearchTool < OpenAI::Internal::Type::BaseModel # The type of the file search tool. Always `file_search`. sig { returns(Symbol) } attr_accessor :type @@ -18,7 +18,7 @@ module OpenAI sig do params( - filters: T.any(OpenAI::Models::ComparisonFilter, OpenAI::Internal::Util::AnyHash, OpenAI::Models::CompoundFilter) + filters: T.any(OpenAI::Models::ComparisonFilter, OpenAI::Internal::AnyHash, OpenAI::Models::CompoundFilter) ) .void end @@ -38,7 +38,7 @@ module OpenAI sig do params( - ranking_options: T.any(OpenAI::Models::Responses::FileSearchTool::RankingOptions, OpenAI::Internal::Util::AnyHash) + ranking_options: T.any(OpenAI::Models::Responses::FileSearchTool::RankingOptions, OpenAI::Internal::AnyHash) ) .void end @@ -50,9 +50,9 @@ module OpenAI sig do params( vector_store_ids: T::Array[String], - filters: T.any(OpenAI::Models::ComparisonFilter, OpenAI::Internal::Util::AnyHash, OpenAI::Models::CompoundFilter), + filters: T.any(OpenAI::Models::ComparisonFilter, OpenAI::Internal::AnyHash, OpenAI::Models::CompoundFilter), max_num_results: Integer, - ranking_options: T.any(OpenAI::Models::Responses::FileSearchTool::RankingOptions, OpenAI::Internal::Util::AnyHash), + ranking_options: T.any(OpenAI::Models::Responses::FileSearchTool::RankingOptions, OpenAI::Internal::AnyHash), type: Symbol ) .returns(T.attached_class) @@ -83,14 +83,14 @@ module OpenAI # A filter to apply based on file attributes. module Filters - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([OpenAI::Models::ComparisonFilter, OpenAI::Models::CompoundFilter]) } def self.variants end end - class RankingOptions < OpenAI::BaseModel + class RankingOptions < OpenAI::Internal::Type::BaseModel # The ranker to use for the file search. sig { returns(T.nilable(OpenAI::Models::Responses::FileSearchTool::RankingOptions::Ranker::OrSymbol)) } attr_reader :ranker @@ -129,7 +129,7 @@ module OpenAI # The ranker to use for the file search. module Ranker - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::FileSearchTool::RankingOptions::Ranker) } diff --git a/rbi/lib/openai/models/responses/function_tool.rbi b/rbi/lib/openai/models/responses/function_tool.rbi index 7462dada..e329cc11 100644 --- a/rbi/lib/openai/models/responses/function_tool.rbi +++ b/rbi/lib/openai/models/responses/function_tool.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class FunctionTool < OpenAI::BaseModel + class FunctionTool < OpenAI::Internal::Type::BaseModel # The name of the function to call. sig { returns(String) } attr_accessor :name diff --git a/rbi/lib/openai/models/responses/input_item_list_params.rbi b/rbi/lib/openai/models/responses/input_item_list_params.rbi index 20120ad6..baf516d6 100644 --- a/rbi/lib/openai/models/responses/input_item_list_params.rbi +++ b/rbi/lib/openai/models/responses/input_item_list_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class InputItemListParams < OpenAI::BaseModel + class InputItemListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -54,7 +54,7 @@ module OpenAI include: T::Array[OpenAI::Models::Responses::ResponseIncludable::OrSymbol], limit: Integer, order: OpenAI::Models::Responses::InputItemListParams::Order::OrSymbol, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -82,7 +82,7 @@ module OpenAI # - `asc`: Return the input items in ascending order. # - `desc`: Return the input items in descending order. module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::InputItemListParams::Order) } OrSymbol = diff --git a/rbi/lib/openai/models/responses/response.rbi b/rbi/lib/openai/models/responses/response.rbi index c2c8686f..4a3056e9 100644 --- a/rbi/lib/openai/models/responses/response.rbi +++ b/rbi/lib/openai/models/responses/response.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class Response < OpenAI::BaseModel + class Response < OpenAI::Internal::Type::BaseModel # Unique identifier for this Response. sig { returns(String) } attr_accessor :id @@ -16,10 +16,7 @@ module OpenAI sig { returns(T.nilable(OpenAI::Models::Responses::ResponseError)) } attr_reader :error - sig do - params(error: T.nilable(T.any(OpenAI::Models::Responses::ResponseError, OpenAI::Internal::Util::AnyHash))) - .void - end + sig { params(error: T.nilable(T.any(OpenAI::Models::Responses::ResponseError, OpenAI::Internal::AnyHash))).void } attr_writer :error # Details about why the response is incomplete. @@ -28,7 +25,7 @@ module OpenAI sig do params( - incomplete_details: T.nilable(T.any(OpenAI::Models::Responses::Response::IncompleteDetails, OpenAI::Internal::Util::AnyHash)) + incomplete_details: T.nilable(T.any(OpenAI::Models::Responses::Response::IncompleteDetails, OpenAI::Internal::AnyHash)) ) .void end @@ -171,7 +168,7 @@ module OpenAI sig { returns(T.nilable(OpenAI::Models::Reasoning)) } attr_reader :reasoning - sig { params(reasoning: T.nilable(T.any(OpenAI::Models::Reasoning, OpenAI::Internal::Util::AnyHash))).void } + sig { params(reasoning: T.nilable(T.any(OpenAI::Models::Reasoning, OpenAI::Internal::AnyHash))).void } attr_writer :reasoning # The status of the response generation. One of `completed`, `failed`, @@ -190,7 +187,7 @@ module OpenAI sig { returns(T.nilable(OpenAI::Models::Responses::ResponseTextConfig)) } attr_reader :text - sig { params(text: T.any(OpenAI::Models::Responses::ResponseTextConfig, OpenAI::Internal::Util::AnyHash)).void } + sig { params(text: T.any(OpenAI::Models::Responses::ResponseTextConfig, OpenAI::Internal::AnyHash)).void } attr_writer :text # The truncation strategy to use for the model response. @@ -208,7 +205,7 @@ module OpenAI sig { returns(T.nilable(OpenAI::Models::Responses::ResponseUsage)) } attr_reader :usage - sig { params(usage: T.any(OpenAI::Models::Responses::ResponseUsage, OpenAI::Internal::Util::AnyHash)).void } + sig { params(usage: T.any(OpenAI::Models::Responses::ResponseUsage, OpenAI::Internal::AnyHash)).void } attr_writer :usage # A unique identifier representing your end-user, which can help OpenAI to monitor @@ -224,15 +221,15 @@ module OpenAI params( id: String, created_at: Float, - error: T.nilable(T.any(OpenAI::Models::Responses::ResponseError, OpenAI::Internal::Util::AnyHash)), - incomplete_details: T.nilable(T.any(OpenAI::Models::Responses::Response::IncompleteDetails, OpenAI::Internal::Util::AnyHash)), + error: T.nilable(T.any(OpenAI::Models::Responses::ResponseError, OpenAI::Internal::AnyHash)), + incomplete_details: T.nilable(T.any(OpenAI::Models::Responses::Response::IncompleteDetails, OpenAI::Internal::AnyHash)), instructions: T.nilable(String), metadata: T.nilable(T::Hash[Symbol, String]), model: T.any(String, OpenAI::Models::ChatModel::OrSymbol, OpenAI::Models::ResponsesModel::OrSymbol), output: T::Array[ T.any( OpenAI::Models::Responses::ResponseOutputMessage, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ResponseFileSearchToolCall, OpenAI::Models::Responses::ResponseFunctionToolCall, OpenAI::Models::Responses::ResponseFunctionWebSearch, @@ -245,13 +242,13 @@ module OpenAI tool_choice: T.any( OpenAI::Models::Responses::ToolChoiceOptions::OrSymbol, OpenAI::Models::Responses::ToolChoiceTypes, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ToolChoiceFunction ), tools: T::Array[ T.any( OpenAI::Models::Responses::FileSearchTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::FunctionTool, OpenAI::Models::Responses::ComputerTool, OpenAI::Models::Responses::WebSearchTool @@ -260,11 +257,11 @@ module OpenAI top_p: T.nilable(Float), max_output_tokens: T.nilable(Integer), previous_response_id: T.nilable(String), - reasoning: T.nilable(T.any(OpenAI::Models::Reasoning, OpenAI::Internal::Util::AnyHash)), + reasoning: T.nilable(T.any(OpenAI::Models::Reasoning, OpenAI::Internal::AnyHash)), status: OpenAI::Models::Responses::ResponseStatus::OrSymbol, - text: T.any(OpenAI::Models::Responses::ResponseTextConfig, OpenAI::Internal::Util::AnyHash), + text: T.any(OpenAI::Models::Responses::ResponseTextConfig, OpenAI::Internal::AnyHash), truncation: T.nilable(OpenAI::Models::Responses::Response::Truncation::OrSymbol), - usage: T.any(OpenAI::Models::Responses::ResponseUsage, OpenAI::Internal::Util::AnyHash), + usage: T.any(OpenAI::Models::Responses::ResponseUsage, OpenAI::Internal::AnyHash), user: String, object: Symbol ) @@ -348,7 +345,7 @@ module OpenAI def to_hash end - class IncompleteDetails < OpenAI::BaseModel + class IncompleteDetails < OpenAI::Internal::Type::BaseModel # The reason why the response is incomplete. sig { returns(T.nilable(OpenAI::Models::Responses::Response::IncompleteDetails::Reason::TaggedSymbol)) } attr_reader :reason @@ -370,7 +367,7 @@ module OpenAI # The reason why the response is incomplete. module Reason - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::Response::IncompleteDetails::Reason) } @@ -392,7 +389,7 @@ module OpenAI # response. See the `tools` parameter to see how to specify which tools the model # can call. module ToolChoice - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override @@ -412,7 +409,7 @@ module OpenAI # - `disabled` (default): If a model response will exceed the context window size # for a model, the request will fail with a 400 error. module Truncation - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::Response::Truncation) } OrSymbol = diff --git a/rbi/lib/openai/models/responses/response_audio_delta_event.rbi b/rbi/lib/openai/models/responses/response_audio_delta_event.rbi index af937de8..17fa0ac5 100644 --- a/rbi/lib/openai/models/responses/response_audio_delta_event.rbi +++ b/rbi/lib/openai/models/responses/response_audio_delta_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseAudioDeltaEvent < OpenAI::BaseModel + class ResponseAudioDeltaEvent < OpenAI::Internal::Type::BaseModel # A chunk of Base64 encoded response audio bytes. sig { returns(String) } attr_accessor :delta diff --git a/rbi/lib/openai/models/responses/response_audio_done_event.rbi b/rbi/lib/openai/models/responses/response_audio_done_event.rbi index 2f302c28..b25bf971 100644 --- a/rbi/lib/openai/models/responses/response_audio_done_event.rbi +++ b/rbi/lib/openai/models/responses/response_audio_done_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseAudioDoneEvent < OpenAI::BaseModel + class ResponseAudioDoneEvent < OpenAI::Internal::Type::BaseModel # The type of the event. Always `response.audio.done`. sig { returns(Symbol) } attr_accessor :type diff --git a/rbi/lib/openai/models/responses/response_audio_transcript_delta_event.rbi b/rbi/lib/openai/models/responses/response_audio_transcript_delta_event.rbi index f2206761..8c6021af 100644 --- a/rbi/lib/openai/models/responses/response_audio_transcript_delta_event.rbi +++ b/rbi/lib/openai/models/responses/response_audio_transcript_delta_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseAudioTranscriptDeltaEvent < OpenAI::BaseModel + class ResponseAudioTranscriptDeltaEvent < OpenAI::Internal::Type::BaseModel # The partial transcript of the audio response. sig { returns(String) } attr_accessor :delta diff --git a/rbi/lib/openai/models/responses/response_audio_transcript_done_event.rbi b/rbi/lib/openai/models/responses/response_audio_transcript_done_event.rbi index cb69a888..52a596c6 100644 --- a/rbi/lib/openai/models/responses/response_audio_transcript_done_event.rbi +++ b/rbi/lib/openai/models/responses/response_audio_transcript_done_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseAudioTranscriptDoneEvent < OpenAI::BaseModel + class ResponseAudioTranscriptDoneEvent < OpenAI::Internal::Type::BaseModel # The type of the event. Always `response.audio.transcript.done`. sig { returns(Symbol) } attr_accessor :type diff --git a/rbi/lib/openai/models/responses/response_code_interpreter_call_code_delta_event.rbi b/rbi/lib/openai/models/responses/response_code_interpreter_call_code_delta_event.rbi index 26d256d8..4c8bd0e9 100644 --- a/rbi/lib/openai/models/responses/response_code_interpreter_call_code_delta_event.rbi +++ b/rbi/lib/openai/models/responses/response_code_interpreter_call_code_delta_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseCodeInterpreterCallCodeDeltaEvent < OpenAI::BaseModel + class ResponseCodeInterpreterCallCodeDeltaEvent < OpenAI::Internal::Type::BaseModel # The partial code snippet added by the code interpreter. sig { returns(String) } attr_accessor :delta diff --git a/rbi/lib/openai/models/responses/response_code_interpreter_call_code_done_event.rbi b/rbi/lib/openai/models/responses/response_code_interpreter_call_code_done_event.rbi index c1512e6a..d1ec38b7 100644 --- a/rbi/lib/openai/models/responses/response_code_interpreter_call_code_done_event.rbi +++ b/rbi/lib/openai/models/responses/response_code_interpreter_call_code_done_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseCodeInterpreterCallCodeDoneEvent < OpenAI::BaseModel + class ResponseCodeInterpreterCallCodeDoneEvent < OpenAI::Internal::Type::BaseModel # The final code snippet output by the code interpreter. sig { returns(String) } attr_accessor :code diff --git a/rbi/lib/openai/models/responses/response_code_interpreter_call_completed_event.rbi b/rbi/lib/openai/models/responses/response_code_interpreter_call_completed_event.rbi index d5e6e2f9..5d1c5965 100644 --- a/rbi/lib/openai/models/responses/response_code_interpreter_call_completed_event.rbi +++ b/rbi/lib/openai/models/responses/response_code_interpreter_call_completed_event.rbi @@ -3,14 +3,14 @@ module OpenAI module Models module Responses - class ResponseCodeInterpreterCallCompletedEvent < OpenAI::BaseModel + class ResponseCodeInterpreterCallCompletedEvent < OpenAI::Internal::Type::BaseModel # A tool call to run code. sig { returns(OpenAI::Models::Responses::ResponseCodeInterpreterToolCall) } attr_reader :code_interpreter_call sig do params( - code_interpreter_call: T.any(OpenAI::Models::Responses::ResponseCodeInterpreterToolCall, OpenAI::Internal::Util::AnyHash) + code_interpreter_call: T.any(OpenAI::Models::Responses::ResponseCodeInterpreterToolCall, OpenAI::Internal::AnyHash) ) .void end @@ -27,7 +27,7 @@ module OpenAI # Emitted when the code interpreter call is completed. sig do params( - code_interpreter_call: T.any(OpenAI::Models::Responses::ResponseCodeInterpreterToolCall, OpenAI::Internal::Util::AnyHash), + code_interpreter_call: T.any(OpenAI::Models::Responses::ResponseCodeInterpreterToolCall, OpenAI::Internal::AnyHash), output_index: Integer, type: Symbol ) diff --git a/rbi/lib/openai/models/responses/response_code_interpreter_call_in_progress_event.rbi b/rbi/lib/openai/models/responses/response_code_interpreter_call_in_progress_event.rbi index 5d0d90a1..02d423ad 100644 --- a/rbi/lib/openai/models/responses/response_code_interpreter_call_in_progress_event.rbi +++ b/rbi/lib/openai/models/responses/response_code_interpreter_call_in_progress_event.rbi @@ -3,14 +3,14 @@ module OpenAI module Models module Responses - class ResponseCodeInterpreterCallInProgressEvent < OpenAI::BaseModel + class ResponseCodeInterpreterCallInProgressEvent < OpenAI::Internal::Type::BaseModel # A tool call to run code. sig { returns(OpenAI::Models::Responses::ResponseCodeInterpreterToolCall) } attr_reader :code_interpreter_call sig do params( - code_interpreter_call: T.any(OpenAI::Models::Responses::ResponseCodeInterpreterToolCall, OpenAI::Internal::Util::AnyHash) + code_interpreter_call: T.any(OpenAI::Models::Responses::ResponseCodeInterpreterToolCall, OpenAI::Internal::AnyHash) ) .void end @@ -27,7 +27,7 @@ module OpenAI # Emitted when a code interpreter call is in progress. sig do params( - code_interpreter_call: T.any(OpenAI::Models::Responses::ResponseCodeInterpreterToolCall, OpenAI::Internal::Util::AnyHash), + code_interpreter_call: T.any(OpenAI::Models::Responses::ResponseCodeInterpreterToolCall, OpenAI::Internal::AnyHash), output_index: Integer, type: Symbol ) diff --git a/rbi/lib/openai/models/responses/response_code_interpreter_call_interpreting_event.rbi b/rbi/lib/openai/models/responses/response_code_interpreter_call_interpreting_event.rbi index f2f0f442..47f2b1a2 100644 --- a/rbi/lib/openai/models/responses/response_code_interpreter_call_interpreting_event.rbi +++ b/rbi/lib/openai/models/responses/response_code_interpreter_call_interpreting_event.rbi @@ -3,14 +3,14 @@ module OpenAI module Models module Responses - class ResponseCodeInterpreterCallInterpretingEvent < OpenAI::BaseModel + class ResponseCodeInterpreterCallInterpretingEvent < OpenAI::Internal::Type::BaseModel # A tool call to run code. sig { returns(OpenAI::Models::Responses::ResponseCodeInterpreterToolCall) } attr_reader :code_interpreter_call sig do params( - code_interpreter_call: T.any(OpenAI::Models::Responses::ResponseCodeInterpreterToolCall, OpenAI::Internal::Util::AnyHash) + code_interpreter_call: T.any(OpenAI::Models::Responses::ResponseCodeInterpreterToolCall, OpenAI::Internal::AnyHash) ) .void end @@ -27,7 +27,7 @@ module OpenAI # Emitted when the code interpreter is actively interpreting the code snippet. sig do params( - code_interpreter_call: T.any(OpenAI::Models::Responses::ResponseCodeInterpreterToolCall, OpenAI::Internal::Util::AnyHash), + code_interpreter_call: T.any(OpenAI::Models::Responses::ResponseCodeInterpreterToolCall, OpenAI::Internal::AnyHash), output_index: Integer, type: Symbol ) diff --git a/rbi/lib/openai/models/responses/response_code_interpreter_tool_call.rbi b/rbi/lib/openai/models/responses/response_code_interpreter_tool_call.rbi index 2f23866c..a6157f98 100644 --- a/rbi/lib/openai/models/responses/response_code_interpreter_tool_call.rbi +++ b/rbi/lib/openai/models/responses/response_code_interpreter_tool_call.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseCodeInterpreterToolCall < OpenAI::BaseModel + class ResponseCodeInterpreterToolCall < OpenAI::Internal::Type::BaseModel # The unique ID of the code interpreter tool call. sig { returns(String) } attr_accessor :id @@ -41,7 +41,7 @@ module OpenAI results: T::Array[ T.any( OpenAI::Models::Responses::ResponseCodeInterpreterToolCall::Result::Logs, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ResponseCodeInterpreterToolCall::Result::Files ) ], @@ -75,9 +75,9 @@ module OpenAI # The output of a code interpreter tool call that is text. module Result - extend OpenAI::Union + extend OpenAI::Internal::Type::Union - class Logs < OpenAI::BaseModel + class Logs < OpenAI::Internal::Type::BaseModel # The logs of the code interpreter tool call. sig { returns(String) } attr_accessor :logs @@ -96,7 +96,7 @@ module OpenAI end end - class Files < OpenAI::BaseModel + class Files < OpenAI::Internal::Type::BaseModel sig { returns(T::Array[OpenAI::Models::Responses::ResponseCodeInterpreterToolCall::Result::Files::File]) } attr_accessor :files @@ -110,7 +110,7 @@ module OpenAI files: T::Array[ T.any( OpenAI::Models::Responses::ResponseCodeInterpreterToolCall::Result::Files::File, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ], type: Symbol @@ -132,7 +132,7 @@ module OpenAI def to_hash end - class File < OpenAI::BaseModel + class File < OpenAI::Internal::Type::BaseModel # The ID of the file. sig { returns(String) } attr_accessor :file_id @@ -163,7 +163,7 @@ module OpenAI # The status of the code interpreter tool call. module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseCodeInterpreterToolCall::Status) } diff --git a/rbi/lib/openai/models/responses/response_completed_event.rbi b/rbi/lib/openai/models/responses/response_completed_event.rbi index 5ea9fdfe..c7774329 100644 --- a/rbi/lib/openai/models/responses/response_completed_event.rbi +++ b/rbi/lib/openai/models/responses/response_completed_event.rbi @@ -3,12 +3,12 @@ module OpenAI module Models module Responses - class ResponseCompletedEvent < OpenAI::BaseModel + class ResponseCompletedEvent < OpenAI::Internal::Type::BaseModel # Properties of the completed response. sig { returns(OpenAI::Models::Responses::Response) } attr_reader :response - sig { params(response: T.any(OpenAI::Models::Responses::Response, OpenAI::Internal::Util::AnyHash)).void } + sig { params(response: T.any(OpenAI::Models::Responses::Response, OpenAI::Internal::AnyHash)).void } attr_writer :response # The type of the event. Always `response.completed`. @@ -17,10 +17,7 @@ module OpenAI # Emitted when the model response is complete. sig do - params( - response: T.any(OpenAI::Models::Responses::Response, OpenAI::Internal::Util::AnyHash), - type: Symbol - ) + params(response: T.any(OpenAI::Models::Responses::Response, OpenAI::Internal::AnyHash), type: Symbol) .returns(T.attached_class) end def self.new(response:, type: :"response.completed") diff --git a/rbi/lib/openai/models/responses/response_computer_tool_call.rbi b/rbi/lib/openai/models/responses/response_computer_tool_call.rbi index 18ad0198..8ced5e39 100644 --- a/rbi/lib/openai/models/responses/response_computer_tool_call.rbi +++ b/rbi/lib/openai/models/responses/response_computer_tool_call.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseComputerToolCall < OpenAI::BaseModel + class ResponseComputerToolCall < OpenAI::Internal::Type::BaseModel # The unique ID of the computer call. sig { returns(String) } attr_accessor :id @@ -51,7 +51,7 @@ module OpenAI id: String, action: T.any( OpenAI::Models::Responses::ResponseComputerToolCall::Action::Click, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ResponseComputerToolCall::Action::DoubleClick, OpenAI::Models::Responses::ResponseComputerToolCall::Action::Drag, OpenAI::Models::Responses::ResponseComputerToolCall::Action::Keypress, @@ -62,12 +62,7 @@ module OpenAI OpenAI::Models::Responses::ResponseComputerToolCall::Action::Wait ), call_id: String, - pending_safety_checks: T::Array[ - T.any( - OpenAI::Models::Responses::ResponseComputerToolCall::PendingSafetyCheck, - OpenAI::Internal::Util::AnyHash - ) - ], + pending_safety_checks: T::Array[T.any(OpenAI::Models::Responses::ResponseComputerToolCall::PendingSafetyCheck, OpenAI::Internal::AnyHash)], status: OpenAI::Models::Responses::ResponseComputerToolCall::Status::OrSymbol, type: OpenAI::Models::Responses::ResponseComputerToolCall::Type::OrSymbol ) @@ -104,9 +99,9 @@ module OpenAI # A click action. module Action - extend OpenAI::Union + extend OpenAI::Internal::Type::Union - class Click < OpenAI::BaseModel + class Click < OpenAI::Internal::Type::BaseModel # Indicates which mouse button was pressed during the click. One of `left`, # `right`, `wheel`, `back`, or `forward`. sig { returns(OpenAI::Models::Responses::ResponseComputerToolCall::Action::Click::Button::OrSymbol) } @@ -155,7 +150,7 @@ module OpenAI # Indicates which mouse button was pressed during the click. One of `left`, # `right`, `wheel`, `back`, or `forward`. module Button - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseComputerToolCall::Action::Click::Button) } @@ -190,7 +185,7 @@ module OpenAI end end - class DoubleClick < OpenAI::BaseModel + class DoubleClick < OpenAI::Internal::Type::BaseModel # Specifies the event type. For a double click action, this property is always set # to `double_click`. sig { returns(Symbol) } @@ -214,7 +209,7 @@ module OpenAI end end - class Drag < OpenAI::BaseModel + class Drag < OpenAI::Internal::Type::BaseModel # An array of coordinates representing the path of the drag action. Coordinates # will appear as an array of objects, eg # @@ -235,12 +230,7 @@ module OpenAI # A drag action. sig do params( - path: T::Array[ - T.any( - OpenAI::Models::Responses::ResponseComputerToolCall::Action::Drag::Path, - OpenAI::Internal::Util::AnyHash - ) - ], + path: T::Array[T.any(OpenAI::Models::Responses::ResponseComputerToolCall::Action::Drag::Path, OpenAI::Internal::AnyHash)], type: Symbol ) .returns(T.attached_class) @@ -257,7 +247,7 @@ module OpenAI def to_hash end - class Path < OpenAI::BaseModel + class Path < OpenAI::Internal::Type::BaseModel # The x-coordinate. sig { returns(Integer) } attr_accessor :x @@ -277,7 +267,7 @@ module OpenAI end end - class Keypress < OpenAI::BaseModel + class Keypress < OpenAI::Internal::Type::BaseModel # The combination of keys the model is requesting to be pressed. This is an array # of strings, each representing a key. sig { returns(T::Array[String]) } @@ -298,7 +288,7 @@ module OpenAI end end - class Move < OpenAI::BaseModel + class Move < OpenAI::Internal::Type::BaseModel # Specifies the event type. For a move action, this property is always set to # `move`. sig { returns(Symbol) } @@ -322,7 +312,7 @@ module OpenAI end end - class Screenshot < OpenAI::BaseModel + class Screenshot < OpenAI::Internal::Type::BaseModel # Specifies the event type. For a screenshot action, this property is always set # to `screenshot`. sig { returns(Symbol) } @@ -338,7 +328,7 @@ module OpenAI end end - class Scroll < OpenAI::BaseModel + class Scroll < OpenAI::Internal::Type::BaseModel # The horizontal scroll distance. sig { returns(Integer) } attr_accessor :scroll_x @@ -375,7 +365,7 @@ module OpenAI end end - class Type < OpenAI::BaseModel + class Type < OpenAI::Internal::Type::BaseModel # The text to type. sig { returns(String) } attr_accessor :text @@ -395,7 +385,7 @@ module OpenAI end end - class Wait < OpenAI::BaseModel + class Wait < OpenAI::Internal::Type::BaseModel # Specifies the event type. For a wait action, this property is always set to # `wait`. sig { returns(Symbol) } @@ -421,7 +411,7 @@ module OpenAI end end - class PendingSafetyCheck < OpenAI::BaseModel + class PendingSafetyCheck < OpenAI::Internal::Type::BaseModel # The ID of the pending safety check. sig { returns(String) } attr_accessor :id @@ -447,7 +437,7 @@ module OpenAI # The status of the item. One of `in_progress`, `completed`, or `incomplete`. # Populated when items are returned via API. module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseComputerToolCall::Status) } @@ -467,7 +457,7 @@ module OpenAI # The type of the computer call. Always `computer_call`. module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseComputerToolCall::Type) } OrSymbol = diff --git a/rbi/lib/openai/models/responses/response_computer_tool_call_output_item.rbi b/rbi/lib/openai/models/responses/response_computer_tool_call_output_item.rbi index 5f0a22d5..fbfa11b5 100644 --- a/rbi/lib/openai/models/responses/response_computer_tool_call_output_item.rbi +++ b/rbi/lib/openai/models/responses/response_computer_tool_call_output_item.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseComputerToolCallOutputItem < OpenAI::BaseModel + class ResponseComputerToolCallOutputItem < OpenAI::Internal::Type::BaseModel # The unique ID of the computer call tool output. sig { returns(String) } attr_accessor :id @@ -18,10 +18,7 @@ module OpenAI sig do params( - output: T.any( - OpenAI::Models::Responses::ResponseComputerToolCallOutputScreenshot, - OpenAI::Internal::Util::AnyHash - ) + output: T.any(OpenAI::Models::Responses::ResponseComputerToolCallOutputScreenshot, OpenAI::Internal::AnyHash) ) .void end @@ -47,7 +44,7 @@ module OpenAI acknowledged_safety_checks: T::Array[ T.any( OpenAI::Models::Responses::ResponseComputerToolCallOutputItem::AcknowledgedSafetyCheck, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ] ) @@ -67,14 +64,11 @@ module OpenAI params( id: String, call_id: String, - output: T.any( - OpenAI::Models::Responses::ResponseComputerToolCallOutputScreenshot, - OpenAI::Internal::Util::AnyHash - ), + output: T.any(OpenAI::Models::Responses::ResponseComputerToolCallOutputScreenshot, OpenAI::Internal::AnyHash), acknowledged_safety_checks: T::Array[ T.any( OpenAI::Models::Responses::ResponseComputerToolCallOutputItem::AcknowledgedSafetyCheck, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ], status: OpenAI::Models::Responses::ResponseComputerToolCallOutputItem::Status::OrSymbol, @@ -108,7 +102,7 @@ module OpenAI def to_hash end - class AcknowledgedSafetyCheck < OpenAI::BaseModel + class AcknowledgedSafetyCheck < OpenAI::Internal::Type::BaseModel # The ID of the pending safety check. sig { returns(String) } attr_accessor :id @@ -134,7 +128,7 @@ module OpenAI # The status of the message input. One of `in_progress`, `completed`, or # `incomplete`. Populated when input items are returned via API. module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseComputerToolCallOutputItem::Status) } diff --git a/rbi/lib/openai/models/responses/response_computer_tool_call_output_screenshot.rbi b/rbi/lib/openai/models/responses/response_computer_tool_call_output_screenshot.rbi index 1246a3c0..11d1b8d6 100644 --- a/rbi/lib/openai/models/responses/response_computer_tool_call_output_screenshot.rbi +++ b/rbi/lib/openai/models/responses/response_computer_tool_call_output_screenshot.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseComputerToolCallOutputScreenshot < OpenAI::BaseModel + class ResponseComputerToolCallOutputScreenshot < OpenAI::Internal::Type::BaseModel # Specifies the event type. For a computer screenshot, this property is always set # to `computer_screenshot`. sig { returns(Symbol) } diff --git a/rbi/lib/openai/models/responses/response_content.rbi b/rbi/lib/openai/models/responses/response_content.rbi index ef91ffdf..5bb9113e 100644 --- a/rbi/lib/openai/models/responses/response_content.rbi +++ b/rbi/lib/openai/models/responses/response_content.rbi @@ -5,7 +5,7 @@ module OpenAI module Responses # Multi-modal input and output contents. module ResponseContent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/responses/response_content_part_added_event.rbi b/rbi/lib/openai/models/responses/response_content_part_added_event.rbi index 9065af5e..669ce321 100644 --- a/rbi/lib/openai/models/responses/response_content_part_added_event.rbi +++ b/rbi/lib/openai/models/responses/response_content_part_added_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseContentPartAddedEvent < OpenAI::BaseModel + class ResponseContentPartAddedEvent < OpenAI::Internal::Type::BaseModel # The index of the content part that was added. sig { returns(Integer) } attr_accessor :content_index @@ -36,7 +36,7 @@ module OpenAI output_index: Integer, part: T.any( OpenAI::Models::Responses::ResponseOutputText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ResponseOutputRefusal ), type: Symbol @@ -63,7 +63,7 @@ module OpenAI # The content part that was added. module Part - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/responses/response_content_part_done_event.rbi b/rbi/lib/openai/models/responses/response_content_part_done_event.rbi index 1b8ece5c..7d1b7312 100644 --- a/rbi/lib/openai/models/responses/response_content_part_done_event.rbi +++ b/rbi/lib/openai/models/responses/response_content_part_done_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseContentPartDoneEvent < OpenAI::BaseModel + class ResponseContentPartDoneEvent < OpenAI::Internal::Type::BaseModel # The index of the content part that is done. sig { returns(Integer) } attr_accessor :content_index @@ -36,7 +36,7 @@ module OpenAI output_index: Integer, part: T.any( OpenAI::Models::Responses::ResponseOutputText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ResponseOutputRefusal ), type: Symbol @@ -63,7 +63,7 @@ module OpenAI # The content part that is done. module Part - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/responses/response_create_params.rbi b/rbi/lib/openai/models/responses/response_create_params.rbi index ef7c1ca5..5fe13ae5 100644 --- a/rbi/lib/openai/models/responses/response_create_params.rbi +++ b/rbi/lib/openai/models/responses/response_create_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseCreateParams < OpenAI::BaseModel + class ResponseCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -100,7 +100,7 @@ module OpenAI sig { returns(T.nilable(OpenAI::Models::Reasoning)) } attr_reader :reasoning - sig { params(reasoning: T.nilable(T.any(OpenAI::Models::Reasoning, OpenAI::Internal::Util::AnyHash))).void } + sig { params(reasoning: T.nilable(T.any(OpenAI::Models::Reasoning, OpenAI::Internal::AnyHash))).void } attr_writer :reasoning # Whether to store the generated model response for later retrieval via API. @@ -122,7 +122,7 @@ module OpenAI sig { returns(T.nilable(OpenAI::Models::Responses::ResponseTextConfig)) } attr_reader :text - sig { params(text: T.any(OpenAI::Models::Responses::ResponseTextConfig, OpenAI::Internal::Util::AnyHash)).void } + sig { params(text: T.any(OpenAI::Models::Responses::ResponseTextConfig, OpenAI::Internal::AnyHash)).void } attr_writer :text # How the model should select which tool (or tools) to use when generating a @@ -146,7 +146,7 @@ module OpenAI tool_choice: T.any( OpenAI::Models::Responses::ToolChoiceOptions::OrSymbol, OpenAI::Models::Responses::ToolChoiceTypes, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ToolChoiceFunction ) ) @@ -189,7 +189,7 @@ module OpenAI tools: T::Array[ T.any( OpenAI::Models::Responses::FileSearchTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::FunctionTool, OpenAI::Models::Responses::ComputerTool, OpenAI::Models::Responses::WebSearchTool @@ -234,7 +234,7 @@ module OpenAI T::Array[ T.any( OpenAI::Models::Responses::EasyInputMessage, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ResponseInputItem::Message, OpenAI::Models::Responses::ResponseOutputMessage, OpenAI::Models::Responses::ResponseFileSearchToolCall, @@ -255,20 +255,20 @@ module OpenAI metadata: T.nilable(T::Hash[Symbol, String]), parallel_tool_calls: T.nilable(T::Boolean), previous_response_id: T.nilable(String), - reasoning: T.nilable(T.any(OpenAI::Models::Reasoning, OpenAI::Internal::Util::AnyHash)), + reasoning: T.nilable(T.any(OpenAI::Models::Reasoning, OpenAI::Internal::AnyHash)), store: T.nilable(T::Boolean), temperature: T.nilable(Float), - text: T.any(OpenAI::Models::Responses::ResponseTextConfig, OpenAI::Internal::Util::AnyHash), + text: T.any(OpenAI::Models::Responses::ResponseTextConfig, OpenAI::Internal::AnyHash), tool_choice: T.any( OpenAI::Models::Responses::ToolChoiceOptions::OrSymbol, OpenAI::Models::Responses::ToolChoiceTypes, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ToolChoiceFunction ), tools: T::Array[ T.any( OpenAI::Models::Responses::FileSearchTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::FunctionTool, OpenAI::Models::Responses::ComputerTool, OpenAI::Models::Responses::WebSearchTool @@ -277,7 +277,7 @@ module OpenAI top_p: T.nilable(Float), truncation: T.nilable(OpenAI::Models::Responses::ResponseCreateParams::Truncation::OrSymbol), user: String, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -369,7 +369,7 @@ module OpenAI # - [Conversation state](https://platform.openai.com/docs/guides/conversation-state) # - [Function calling](https://platform.openai.com/docs/guides/function-calling) module Input - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override @@ -402,7 +402,7 @@ module OpenAI # response. See the `tools` parameter to see how to specify which tools the model # can call. module ToolChoice - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override @@ -422,7 +422,7 @@ module OpenAI # - `disabled` (default): If a model response will exceed the context window size # for a model, the request will fail with a 400 error. module Truncation - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseCreateParams::Truncation) } diff --git a/rbi/lib/openai/models/responses/response_created_event.rbi b/rbi/lib/openai/models/responses/response_created_event.rbi index 0ce79c50..ee2e69fc 100644 --- a/rbi/lib/openai/models/responses/response_created_event.rbi +++ b/rbi/lib/openai/models/responses/response_created_event.rbi @@ -3,12 +3,12 @@ module OpenAI module Models module Responses - class ResponseCreatedEvent < OpenAI::BaseModel + class ResponseCreatedEvent < OpenAI::Internal::Type::BaseModel # The response that was created. sig { returns(OpenAI::Models::Responses::Response) } attr_reader :response - sig { params(response: T.any(OpenAI::Models::Responses::Response, OpenAI::Internal::Util::AnyHash)).void } + sig { params(response: T.any(OpenAI::Models::Responses::Response, OpenAI::Internal::AnyHash)).void } attr_writer :response # The type of the event. Always `response.created`. @@ -17,10 +17,7 @@ module OpenAI # An event that is emitted when a response is created. sig do - params( - response: T.any(OpenAI::Models::Responses::Response, OpenAI::Internal::Util::AnyHash), - type: Symbol - ) + params(response: T.any(OpenAI::Models::Responses::Response, OpenAI::Internal::AnyHash), type: Symbol) .returns(T.attached_class) end def self.new(response:, type: :"response.created") diff --git a/rbi/lib/openai/models/responses/response_delete_params.rbi b/rbi/lib/openai/models/responses/response_delete_params.rbi index 008c4bd9..ac0c636e 100644 --- a/rbi/lib/openai/models/responses/response_delete_params.rbi +++ b/rbi/lib/openai/models/responses/response_delete_params.rbi @@ -3,13 +3,17 @@ module OpenAI module Models module Responses - class ResponseDeleteParams < OpenAI::BaseModel + class ResponseDeleteParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters sig do - params(request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) - .returns(T.attached_class) + params( + request_options: T.any( + OpenAI::RequestOptions, + OpenAI::Internal::AnyHash + ) + ).returns(T.attached_class) end def self.new(request_options: {}) end diff --git a/rbi/lib/openai/models/responses/response_error.rbi b/rbi/lib/openai/models/responses/response_error.rbi index 4f682bd4..a6cb9d7e 100644 --- a/rbi/lib/openai/models/responses/response_error.rbi +++ b/rbi/lib/openai/models/responses/response_error.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseError < OpenAI::BaseModel + class ResponseError < OpenAI::Internal::Type::BaseModel # The error code for the response. sig { returns(OpenAI::Models::Responses::ResponseError::Code::TaggedSymbol) } attr_accessor :code @@ -26,7 +26,7 @@ module OpenAI # The error code for the response. module Code - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseError::Code) } OrSymbol = diff --git a/rbi/lib/openai/models/responses/response_error_event.rbi b/rbi/lib/openai/models/responses/response_error_event.rbi index 303a0210..55680bad 100644 --- a/rbi/lib/openai/models/responses/response_error_event.rbi +++ b/rbi/lib/openai/models/responses/response_error_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseErrorEvent < OpenAI::BaseModel + class ResponseErrorEvent < OpenAI::Internal::Type::BaseModel # The error code. sig { returns(T.nilable(String)) } attr_accessor :code diff --git a/rbi/lib/openai/models/responses/response_failed_event.rbi b/rbi/lib/openai/models/responses/response_failed_event.rbi index 8d43369b..fdbaf6f0 100644 --- a/rbi/lib/openai/models/responses/response_failed_event.rbi +++ b/rbi/lib/openai/models/responses/response_failed_event.rbi @@ -3,12 +3,12 @@ module OpenAI module Models module Responses - class ResponseFailedEvent < OpenAI::BaseModel + class ResponseFailedEvent < OpenAI::Internal::Type::BaseModel # The response that failed. sig { returns(OpenAI::Models::Responses::Response) } attr_reader :response - sig { params(response: T.any(OpenAI::Models::Responses::Response, OpenAI::Internal::Util::AnyHash)).void } + sig { params(response: T.any(OpenAI::Models::Responses::Response, OpenAI::Internal::AnyHash)).void } attr_writer :response # The type of the event. Always `response.failed`. @@ -17,10 +17,7 @@ module OpenAI # An event that is emitted when a response fails. sig do - params( - response: T.any(OpenAI::Models::Responses::Response, OpenAI::Internal::Util::AnyHash), - type: Symbol - ) + params(response: T.any(OpenAI::Models::Responses::Response, OpenAI::Internal::AnyHash), type: Symbol) .returns(T.attached_class) end def self.new(response:, type: :"response.failed") diff --git a/rbi/lib/openai/models/responses/response_file_search_call_completed_event.rbi b/rbi/lib/openai/models/responses/response_file_search_call_completed_event.rbi index fd8d4c62..285df33d 100644 --- a/rbi/lib/openai/models/responses/response_file_search_call_completed_event.rbi +++ b/rbi/lib/openai/models/responses/response_file_search_call_completed_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseFileSearchCallCompletedEvent < OpenAI::BaseModel + class ResponseFileSearchCallCompletedEvent < OpenAI::Internal::Type::BaseModel # The ID of the output item that the file search call is initiated. sig { returns(String) } attr_accessor :item_id diff --git a/rbi/lib/openai/models/responses/response_file_search_call_in_progress_event.rbi b/rbi/lib/openai/models/responses/response_file_search_call_in_progress_event.rbi index 621c0706..e6d95bd6 100644 --- a/rbi/lib/openai/models/responses/response_file_search_call_in_progress_event.rbi +++ b/rbi/lib/openai/models/responses/response_file_search_call_in_progress_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseFileSearchCallInProgressEvent < OpenAI::BaseModel + class ResponseFileSearchCallInProgressEvent < OpenAI::Internal::Type::BaseModel # The ID of the output item that the file search call is initiated. sig { returns(String) } attr_accessor :item_id diff --git a/rbi/lib/openai/models/responses/response_file_search_call_searching_event.rbi b/rbi/lib/openai/models/responses/response_file_search_call_searching_event.rbi index 849e6603..a6f01d00 100644 --- a/rbi/lib/openai/models/responses/response_file_search_call_searching_event.rbi +++ b/rbi/lib/openai/models/responses/response_file_search_call_searching_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseFileSearchCallSearchingEvent < OpenAI::BaseModel + class ResponseFileSearchCallSearchingEvent < OpenAI::Internal::Type::BaseModel # The ID of the output item that the file search call is initiated. sig { returns(String) } attr_accessor :item_id diff --git a/rbi/lib/openai/models/responses/response_file_search_tool_call.rbi b/rbi/lib/openai/models/responses/response_file_search_tool_call.rbi index 050fdfc1..98e75a9c 100644 --- a/rbi/lib/openai/models/responses/response_file_search_tool_call.rbi +++ b/rbi/lib/openai/models/responses/response_file_search_tool_call.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseFileSearchToolCall < OpenAI::BaseModel + class ResponseFileSearchToolCall < OpenAI::Internal::Type::BaseModel # The unique ID of the file search tool call. sig { returns(String) } attr_accessor :id @@ -34,7 +34,7 @@ module OpenAI queries: T::Array[String], status: OpenAI::Models::Responses::ResponseFileSearchToolCall::Status::OrSymbol, results: T.nilable( - T::Array[T.any(OpenAI::Models::Responses::ResponseFileSearchToolCall::Result, OpenAI::Internal::Util::AnyHash)] + T::Array[T.any(OpenAI::Models::Responses::ResponseFileSearchToolCall::Result, OpenAI::Internal::AnyHash)] ), type: Symbol ) @@ -61,7 +61,7 @@ module OpenAI # The status of the file search tool call. One of `in_progress`, `searching`, # `incomplete` or `failed`, module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseFileSearchToolCall::Status) } @@ -83,7 +83,7 @@ module OpenAI end end - class Result < OpenAI::BaseModel + class Result < OpenAI::Internal::Type::BaseModel # Set of 16 key-value pairs that can be attached to an object. This can be useful # for storing additional information about the object in a structured format, and # querying for objects via API or the dashboard. Keys are strings with a maximum @@ -149,7 +149,7 @@ module OpenAI end module Attribute - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, Float, T::Boolean]) } def self.variants diff --git a/rbi/lib/openai/models/responses/response_format_text_config.rbi b/rbi/lib/openai/models/responses/response_format_text_config.rbi index 24160f1c..f05a3373 100644 --- a/rbi/lib/openai/models/responses/response_format_text_config.rbi +++ b/rbi/lib/openai/models/responses/response_format_text_config.rbi @@ -17,7 +17,7 @@ module OpenAI # ensures the message the model generates is valid JSON. Using `json_schema` is # preferred for models that support it. module ResponseFormatTextConfig - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/responses/response_format_text_json_schema_config.rbi b/rbi/lib/openai/models/responses/response_format_text_json_schema_config.rbi index 8d657f44..294d5b32 100644 --- a/rbi/lib/openai/models/responses/response_format_text_json_schema_config.rbi +++ b/rbi/lib/openai/models/responses/response_format_text_json_schema_config.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseFormatTextJSONSchemaConfig < OpenAI::BaseModel + class ResponseFormatTextJSONSchemaConfig < OpenAI::Internal::Type::BaseModel # The name of the response format. Must be a-z, A-Z, 0-9, or contain underscores # and dashes, with a maximum length of 64. sig { returns(String) } diff --git a/rbi/lib/openai/models/responses/response_function_call_arguments_delta_event.rbi b/rbi/lib/openai/models/responses/response_function_call_arguments_delta_event.rbi index 5aada226..8d5a0f74 100644 --- a/rbi/lib/openai/models/responses/response_function_call_arguments_delta_event.rbi +++ b/rbi/lib/openai/models/responses/response_function_call_arguments_delta_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseFunctionCallArgumentsDeltaEvent < OpenAI::BaseModel + class ResponseFunctionCallArgumentsDeltaEvent < OpenAI::Internal::Type::BaseModel # The function-call arguments delta that is added. sig { returns(String) } attr_accessor :delta diff --git a/rbi/lib/openai/models/responses/response_function_call_arguments_done_event.rbi b/rbi/lib/openai/models/responses/response_function_call_arguments_done_event.rbi index ba8b13d1..49d5fc83 100644 --- a/rbi/lib/openai/models/responses/response_function_call_arguments_done_event.rbi +++ b/rbi/lib/openai/models/responses/response_function_call_arguments_done_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseFunctionCallArgumentsDoneEvent < OpenAI::BaseModel + class ResponseFunctionCallArgumentsDoneEvent < OpenAI::Internal::Type::BaseModel # The function-call arguments. sig { returns(String) } attr_accessor :arguments diff --git a/rbi/lib/openai/models/responses/response_function_tool_call.rbi b/rbi/lib/openai/models/responses/response_function_tool_call.rbi index f7b0b0a7..32854d1a 100644 --- a/rbi/lib/openai/models/responses/response_function_tool_call.rbi +++ b/rbi/lib/openai/models/responses/response_function_tool_call.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseFunctionToolCall < OpenAI::BaseModel + class ResponseFunctionToolCall < OpenAI::Internal::Type::BaseModel # A JSON string of the arguments to pass to the function. sig { returns(String) } attr_accessor :arguments @@ -71,7 +71,7 @@ module OpenAI # The status of the item. One of `in_progress`, `completed`, or `incomplete`. # Populated when items are returned via API. module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseFunctionToolCall::Status) } diff --git a/rbi/lib/openai/models/responses/response_function_tool_call_output_item.rbi b/rbi/lib/openai/models/responses/response_function_tool_call_output_item.rbi index b6e0956c..f2d23ee3 100644 --- a/rbi/lib/openai/models/responses/response_function_tool_call_output_item.rbi +++ b/rbi/lib/openai/models/responses/response_function_tool_call_output_item.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseFunctionToolCallOutputItem < OpenAI::BaseModel + class ResponseFunctionToolCallOutputItem < OpenAI::Internal::Type::BaseModel # The unique ID of the function call tool output. sig { returns(String) } attr_accessor :id @@ -59,7 +59,7 @@ module OpenAI # The status of the item. One of `in_progress`, `completed`, or `incomplete`. # Populated when items are returned via API. module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseFunctionToolCallOutputItem::Status) } diff --git a/rbi/lib/openai/models/responses/response_function_web_search.rbi b/rbi/lib/openai/models/responses/response_function_web_search.rbi index 080ec4db..b479dcf8 100644 --- a/rbi/lib/openai/models/responses/response_function_web_search.rbi +++ b/rbi/lib/openai/models/responses/response_function_web_search.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseFunctionWebSearch < OpenAI::BaseModel + class ResponseFunctionWebSearch < OpenAI::Internal::Type::BaseModel # The unique ID of the web search tool call. sig { returns(String) } attr_accessor :id @@ -41,7 +41,7 @@ module OpenAI # The status of the web search tool call. module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseFunctionWebSearch::Status) } diff --git a/rbi/lib/openai/models/responses/response_in_progress_event.rbi b/rbi/lib/openai/models/responses/response_in_progress_event.rbi index 21962a06..187ca04d 100644 --- a/rbi/lib/openai/models/responses/response_in_progress_event.rbi +++ b/rbi/lib/openai/models/responses/response_in_progress_event.rbi @@ -3,12 +3,12 @@ module OpenAI module Models module Responses - class ResponseInProgressEvent < OpenAI::BaseModel + class ResponseInProgressEvent < OpenAI::Internal::Type::BaseModel # The response that is in progress. sig { returns(OpenAI::Models::Responses::Response) } attr_reader :response - sig { params(response: T.any(OpenAI::Models::Responses::Response, OpenAI::Internal::Util::AnyHash)).void } + sig { params(response: T.any(OpenAI::Models::Responses::Response, OpenAI::Internal::AnyHash)).void } attr_writer :response # The type of the event. Always `response.in_progress`. @@ -17,10 +17,7 @@ module OpenAI # Emitted when the response is in progress. sig do - params( - response: T.any(OpenAI::Models::Responses::Response, OpenAI::Internal::Util::AnyHash), - type: Symbol - ) + params(response: T.any(OpenAI::Models::Responses::Response, OpenAI::Internal::AnyHash), type: Symbol) .returns(T.attached_class) end def self.new(response:, type: :"response.in_progress") diff --git a/rbi/lib/openai/models/responses/response_includable.rbi b/rbi/lib/openai/models/responses/response_includable.rbi index f2015ad7..de36828b 100644 --- a/rbi/lib/openai/models/responses/response_includable.rbi +++ b/rbi/lib/openai/models/responses/response_includable.rbi @@ -12,7 +12,7 @@ module OpenAI # - `computer_call_output.output.image_url`: Include image urls from the computer # call output. module ResponseIncludable - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseIncludable) } OrSymbol = diff --git a/rbi/lib/openai/models/responses/response_incomplete_event.rbi b/rbi/lib/openai/models/responses/response_incomplete_event.rbi index ab06154d..305b7c65 100644 --- a/rbi/lib/openai/models/responses/response_incomplete_event.rbi +++ b/rbi/lib/openai/models/responses/response_incomplete_event.rbi @@ -3,12 +3,12 @@ module OpenAI module Models module Responses - class ResponseIncompleteEvent < OpenAI::BaseModel + class ResponseIncompleteEvent < OpenAI::Internal::Type::BaseModel # The response that was incomplete. sig { returns(OpenAI::Models::Responses::Response) } attr_reader :response - sig { params(response: T.any(OpenAI::Models::Responses::Response, OpenAI::Internal::Util::AnyHash)).void } + sig { params(response: T.any(OpenAI::Models::Responses::Response, OpenAI::Internal::AnyHash)).void } attr_writer :response # The type of the event. Always `response.incomplete`. @@ -17,10 +17,7 @@ module OpenAI # An event that is emitted when a response finishes as incomplete. sig do - params( - response: T.any(OpenAI::Models::Responses::Response, OpenAI::Internal::Util::AnyHash), - type: Symbol - ) + params(response: T.any(OpenAI::Models::Responses::Response, OpenAI::Internal::AnyHash), type: Symbol) .returns(T.attached_class) end def self.new(response:, type: :"response.incomplete") diff --git a/rbi/lib/openai/models/responses/response_input.rbi b/rbi/lib/openai/models/responses/response_input.rbi index f86dda51..77a9217a 100644 --- a/rbi/lib/openai/models/responses/response_input.rbi +++ b/rbi/lib/openai/models/responses/response_input.rbi @@ -5,7 +5,7 @@ module OpenAI module Responses ResponseInput = T.let( - OpenAI::ArrayOf[union: OpenAI::Models::Responses::ResponseInputItem], + OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Responses::ResponseInputItem], OpenAI::Internal::Type::Converter ) end diff --git a/rbi/lib/openai/models/responses/response_input_audio.rbi b/rbi/lib/openai/models/responses/response_input_audio.rbi index 7135e479..cbcdd1d1 100644 --- a/rbi/lib/openai/models/responses/response_input_audio.rbi +++ b/rbi/lib/openai/models/responses/response_input_audio.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseInputAudio < OpenAI::BaseModel + class ResponseInputAudio < OpenAI::Internal::Type::BaseModel # Base64-encoded audio data. sig { returns(String) } attr_accessor :data @@ -39,7 +39,7 @@ module OpenAI # The format of the audio data. Currently supported formats are `mp3` and `wav`. module Format - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseInputAudio::Format) } OrSymbol = diff --git a/rbi/lib/openai/models/responses/response_input_content.rbi b/rbi/lib/openai/models/responses/response_input_content.rbi index d63620b1..e1a66fb0 100644 --- a/rbi/lib/openai/models/responses/response_input_content.rbi +++ b/rbi/lib/openai/models/responses/response_input_content.rbi @@ -5,7 +5,7 @@ module OpenAI module Responses # A text input to the model. module ResponseInputContent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/responses/response_input_file.rbi b/rbi/lib/openai/models/responses/response_input_file.rbi index 05c60269..81323132 100644 --- a/rbi/lib/openai/models/responses/response_input_file.rbi +++ b/rbi/lib/openai/models/responses/response_input_file.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseInputFile < OpenAI::BaseModel + class ResponseInputFile < OpenAI::Internal::Type::BaseModel # The type of the input item. Always `input_file`. sig { returns(Symbol) } attr_accessor :type diff --git a/rbi/lib/openai/models/responses/response_input_image.rbi b/rbi/lib/openai/models/responses/response_input_image.rbi index 11947564..a15a6bc9 100644 --- a/rbi/lib/openai/models/responses/response_input_image.rbi +++ b/rbi/lib/openai/models/responses/response_input_image.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseInputImage < OpenAI::BaseModel + class ResponseInputImage < OpenAI::Internal::Type::BaseModel # The detail level of the image to be sent to the model. One of `high`, `low`, or # `auto`. Defaults to `auto`. sig { returns(OpenAI::Models::Responses::ResponseInputImage::Detail::OrSymbol) } @@ -53,7 +53,7 @@ module OpenAI # The detail level of the image to be sent to the model. One of `high`, `low`, or # `auto`. Defaults to `auto`. module Detail - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseInputImage::Detail) } OrSymbol = diff --git a/rbi/lib/openai/models/responses/response_input_item.rbi b/rbi/lib/openai/models/responses/response_input_item.rbi index ab326c72..c01e2b4f 100644 --- a/rbi/lib/openai/models/responses/response_input_item.rbi +++ b/rbi/lib/openai/models/responses/response_input_item.rbi @@ -9,9 +9,9 @@ module OpenAI # `assistant` role are presumed to have been generated by the model in previous # interactions. module ResponseInputItem - extend OpenAI::Union + extend OpenAI::Internal::Type::Union - class Message < OpenAI::BaseModel + class Message < OpenAI::Internal::Type::BaseModel # A list of one or many input items to the model, containing different content # types. sig do @@ -54,7 +54,7 @@ module OpenAI content: T::Array[ T.any( OpenAI::Models::Responses::ResponseInputText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ResponseInputImage, OpenAI::Models::Responses::ResponseInputFile ) @@ -90,7 +90,7 @@ module OpenAI # The role of the message input. One of `user`, `system`, or `developer`. module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseInputItem::Message::Role) } @@ -109,7 +109,7 @@ module OpenAI # The status of item. One of `in_progress`, `completed`, or `incomplete`. # Populated when items are returned via API. module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseInputItem::Message::Status) } @@ -130,7 +130,7 @@ module OpenAI # The type of the message input. Always set to `message`. module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseInputItem::Message::Type) } @@ -145,7 +145,7 @@ module OpenAI end end - class ComputerCallOutput < OpenAI::BaseModel + class ComputerCallOutput < OpenAI::Internal::Type::BaseModel # The ID of the computer tool call that produced the output. sig { returns(String) } attr_accessor :call_id @@ -156,10 +156,7 @@ module OpenAI sig do params( - output: T.any( - OpenAI::Models::Responses::ResponseComputerToolCallOutputScreenshot, - OpenAI::Internal::Util::AnyHash - ) + output: T.any(OpenAI::Models::Responses::ResponseComputerToolCallOutputScreenshot, OpenAI::Internal::AnyHash) ) .void end @@ -192,7 +189,7 @@ module OpenAI acknowledged_safety_checks: T::Array[ T.any( OpenAI::Models::Responses::ResponseInputItem::ComputerCallOutput::AcknowledgedSafetyCheck, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ] ) @@ -212,15 +209,12 @@ module OpenAI sig do params( call_id: String, - output: T.any( - OpenAI::Models::Responses::ResponseComputerToolCallOutputScreenshot, - OpenAI::Internal::Util::AnyHash - ), + output: T.any(OpenAI::Models::Responses::ResponseComputerToolCallOutputScreenshot, OpenAI::Internal::AnyHash), id: String, acknowledged_safety_checks: T::Array[ T.any( OpenAI::Models::Responses::ResponseInputItem::ComputerCallOutput::AcknowledgedSafetyCheck, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ], status: OpenAI::Models::Responses::ResponseInputItem::ComputerCallOutput::Status::OrSymbol, @@ -254,7 +248,7 @@ module OpenAI def to_hash end - class AcknowledgedSafetyCheck < OpenAI::BaseModel + class AcknowledgedSafetyCheck < OpenAI::Internal::Type::BaseModel # The ID of the pending safety check. sig { returns(String) } attr_accessor :id @@ -280,7 +274,7 @@ module OpenAI # The status of the message input. One of `in_progress`, `completed`, or # `incomplete`. Populated when input items are returned via API. module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseInputItem::ComputerCallOutput::Status) } @@ -312,7 +306,7 @@ module OpenAI end end - class FunctionCallOutput < OpenAI::BaseModel + class FunctionCallOutput < OpenAI::Internal::Type::BaseModel # The unique ID of the function tool call generated by the model. sig { returns(String) } attr_accessor :call_id @@ -373,7 +367,7 @@ module OpenAI # The status of the item. One of `in_progress`, `completed`, or `incomplete`. # Populated when items are returned via API. module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseInputItem::FunctionCallOutput::Status) } @@ -405,7 +399,7 @@ module OpenAI end end - class ItemReference < OpenAI::BaseModel + class ItemReference < OpenAI::Internal::Type::BaseModel # The ID of the item to reference. sig { returns(String) } attr_accessor :id diff --git a/rbi/lib/openai/models/responses/response_input_message_content_list.rbi b/rbi/lib/openai/models/responses/response_input_message_content_list.rbi index b06ad65b..f25328f7 100644 --- a/rbi/lib/openai/models/responses/response_input_message_content_list.rbi +++ b/rbi/lib/openai/models/responses/response_input_message_content_list.rbi @@ -5,7 +5,7 @@ module OpenAI module Responses ResponseInputMessageContentList = T.let( - OpenAI::ArrayOf[union: OpenAI::Models::Responses::ResponseInputContent], + OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Responses::ResponseInputContent], OpenAI::Internal::Type::Converter ) end diff --git a/rbi/lib/openai/models/responses/response_input_message_item.rbi b/rbi/lib/openai/models/responses/response_input_message_item.rbi index 60b55701..90c6a533 100644 --- a/rbi/lib/openai/models/responses/response_input_message_item.rbi +++ b/rbi/lib/openai/models/responses/response_input_message_item.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseInputMessageItem < OpenAI::BaseModel + class ResponseInputMessageItem < OpenAI::Internal::Type::BaseModel # The unique ID of the message input. sig { returns(String) } attr_accessor :id @@ -48,7 +48,7 @@ module OpenAI content: T::Array[ T.any( OpenAI::Models::Responses::ResponseInputText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ResponseInputImage, OpenAI::Models::Responses::ResponseInputFile ) @@ -85,7 +85,7 @@ module OpenAI # The role of the message input. One of `user`, `system`, or `developer`. module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseInputMessageItem::Role) } OrSymbol = @@ -103,7 +103,7 @@ module OpenAI # The status of item. One of `in_progress`, `completed`, or `incomplete`. # Populated when items are returned via API. module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseInputMessageItem::Status) } @@ -123,7 +123,7 @@ module OpenAI # The type of the message input. Always set to `message`. module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseInputMessageItem::Type) } OrSymbol = diff --git a/rbi/lib/openai/models/responses/response_input_text.rbi b/rbi/lib/openai/models/responses/response_input_text.rbi index 5c0838ec..94dd8553 100644 --- a/rbi/lib/openai/models/responses/response_input_text.rbi +++ b/rbi/lib/openai/models/responses/response_input_text.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseInputText < OpenAI::BaseModel + class ResponseInputText < OpenAI::Internal::Type::BaseModel # The text input to the model. sig { returns(String) } attr_accessor :text diff --git a/rbi/lib/openai/models/responses/response_item.rbi b/rbi/lib/openai/models/responses/response_item.rbi index b57094c5..3a0b1c92 100644 --- a/rbi/lib/openai/models/responses/response_item.rbi +++ b/rbi/lib/openai/models/responses/response_item.rbi @@ -5,7 +5,7 @@ module OpenAI module Responses # Content item used to generate a response. module ResponseItem - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/responses/response_item_list.rbi b/rbi/lib/openai/models/responses/response_item_list.rbi index 93cf118a..5b5f1583 100644 --- a/rbi/lib/openai/models/responses/response_item_list.rbi +++ b/rbi/lib/openai/models/responses/response_item_list.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseItemList < OpenAI::BaseModel + class ResponseItemList < OpenAI::Internal::Type::BaseModel # A list of items used to generate this response. sig do returns( @@ -45,7 +45,7 @@ module OpenAI data: T::Array[ T.any( OpenAI::Models::Responses::ResponseInputMessageItem, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ResponseOutputMessage, OpenAI::Models::Responses::ResponseFileSearchToolCall, OpenAI::Models::Responses::ResponseComputerToolCall, diff --git a/rbi/lib/openai/models/responses/response_output_audio.rbi b/rbi/lib/openai/models/responses/response_output_audio.rbi index c5a26f54..53e9b4c3 100644 --- a/rbi/lib/openai/models/responses/response_output_audio.rbi +++ b/rbi/lib/openai/models/responses/response_output_audio.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseOutputAudio < OpenAI::BaseModel + class ResponseOutputAudio < OpenAI::Internal::Type::BaseModel # Base64-encoded audio data from the model. sig { returns(String) } attr_accessor :data diff --git a/rbi/lib/openai/models/responses/response_output_item.rbi b/rbi/lib/openai/models/responses/response_output_item.rbi index 9d893172..5827fb3d 100644 --- a/rbi/lib/openai/models/responses/response_output_item.rbi +++ b/rbi/lib/openai/models/responses/response_output_item.rbi @@ -5,7 +5,7 @@ module OpenAI module Responses # An output message from the model. module ResponseOutputItem - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/responses/response_output_item_added_event.rbi b/rbi/lib/openai/models/responses/response_output_item_added_event.rbi index 735922c6..3e5daa2e 100644 --- a/rbi/lib/openai/models/responses/response_output_item_added_event.rbi +++ b/rbi/lib/openai/models/responses/response_output_item_added_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseOutputItemAddedEvent < OpenAI::BaseModel + class ResponseOutputItemAddedEvent < OpenAI::Internal::Type::BaseModel # The output item that was added. sig do returns( @@ -32,7 +32,7 @@ module OpenAI params( item: T.any( OpenAI::Models::Responses::ResponseOutputMessage, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ResponseFileSearchToolCall, OpenAI::Models::Responses::ResponseFunctionToolCall, OpenAI::Models::Responses::ResponseFunctionWebSearch, diff --git a/rbi/lib/openai/models/responses/response_output_item_done_event.rbi b/rbi/lib/openai/models/responses/response_output_item_done_event.rbi index 8dde062b..e7625f50 100644 --- a/rbi/lib/openai/models/responses/response_output_item_done_event.rbi +++ b/rbi/lib/openai/models/responses/response_output_item_done_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseOutputItemDoneEvent < OpenAI::BaseModel + class ResponseOutputItemDoneEvent < OpenAI::Internal::Type::BaseModel # The output item that was marked done. sig do returns( @@ -32,7 +32,7 @@ module OpenAI params( item: T.any( OpenAI::Models::Responses::ResponseOutputMessage, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ResponseFileSearchToolCall, OpenAI::Models::Responses::ResponseFunctionToolCall, OpenAI::Models::Responses::ResponseFunctionWebSearch, diff --git a/rbi/lib/openai/models/responses/response_output_message.rbi b/rbi/lib/openai/models/responses/response_output_message.rbi index 8b07428a..0a39a8eb 100644 --- a/rbi/lib/openai/models/responses/response_output_message.rbi +++ b/rbi/lib/openai/models/responses/response_output_message.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseOutputMessage < OpenAI::BaseModel + class ResponseOutputMessage < OpenAI::Internal::Type::BaseModel # The unique ID of the output message. sig { returns(String) } attr_accessor :id @@ -36,7 +36,7 @@ module OpenAI content: T::Array[ T.any( OpenAI::Models::Responses::ResponseOutputText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ResponseOutputRefusal ) ], @@ -66,7 +66,7 @@ module OpenAI # A text output from the model. module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override @@ -79,7 +79,7 @@ module OpenAI # The status of the message input. One of `in_progress`, `completed`, or # `incomplete`. Populated when input items are returned via API. module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseOutputMessage::Status) } OrSymbol = diff --git a/rbi/lib/openai/models/responses/response_output_refusal.rbi b/rbi/lib/openai/models/responses/response_output_refusal.rbi index 5e416cf9..9fff39df 100644 --- a/rbi/lib/openai/models/responses/response_output_refusal.rbi +++ b/rbi/lib/openai/models/responses/response_output_refusal.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseOutputRefusal < OpenAI::BaseModel + class ResponseOutputRefusal < OpenAI::Internal::Type::BaseModel # The refusal explanationfrom the model. sig { returns(String) } attr_accessor :refusal diff --git a/rbi/lib/openai/models/responses/response_output_text.rbi b/rbi/lib/openai/models/responses/response_output_text.rbi index d8aeb7e6..e8092e12 100644 --- a/rbi/lib/openai/models/responses/response_output_text.rbi +++ b/rbi/lib/openai/models/responses/response_output_text.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseOutputText < OpenAI::BaseModel + class ResponseOutputText < OpenAI::Internal::Type::BaseModel # The annotations of the text output. sig do returns( @@ -32,7 +32,7 @@ module OpenAI annotations: T::Array[ T.any( OpenAI::Models::Responses::ResponseOutputText::Annotation::FileCitation, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ResponseOutputText::Annotation::URLCitation, OpenAI::Models::Responses::ResponseOutputText::Annotation::FilePath ) @@ -66,9 +66,9 @@ module OpenAI # A citation to a file. module Annotation - extend OpenAI::Union + extend OpenAI::Internal::Type::Union - class FileCitation < OpenAI::BaseModel + class FileCitation < OpenAI::Internal::Type::BaseModel # The ID of the file. sig { returns(String) } attr_accessor :file_id @@ -91,7 +91,7 @@ module OpenAI end end - class URLCitation < OpenAI::BaseModel + class URLCitation < OpenAI::Internal::Type::BaseModel # The index of the last character of the URL citation in the message. sig { returns(Integer) } attr_accessor :end_index @@ -135,7 +135,7 @@ module OpenAI end end - class FilePath < OpenAI::BaseModel + class FilePath < OpenAI::Internal::Type::BaseModel # The ID of the file. sig { returns(String) } attr_accessor :file_id diff --git a/rbi/lib/openai/models/responses/response_reasoning_item.rbi b/rbi/lib/openai/models/responses/response_reasoning_item.rbi index bd3303b1..3359bd3e 100644 --- a/rbi/lib/openai/models/responses/response_reasoning_item.rbi +++ b/rbi/lib/openai/models/responses/response_reasoning_item.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseReasoningItem < OpenAI::BaseModel + class ResponseReasoningItem < OpenAI::Internal::Type::BaseModel # The unique identifier of the reasoning content. sig { returns(String) } attr_accessor :id @@ -29,7 +29,7 @@ module OpenAI sig do params( id: String, - summary: T::Array[T.any(OpenAI::Models::Responses::ResponseReasoningItem::Summary, OpenAI::Internal::Util::AnyHash)], + summary: T::Array[T.any(OpenAI::Models::Responses::ResponseReasoningItem::Summary, OpenAI::Internal::AnyHash)], status: OpenAI::Models::Responses::ResponseReasoningItem::Status::OrSymbol, type: Symbol ) @@ -52,7 +52,7 @@ module OpenAI def to_hash end - class Summary < OpenAI::BaseModel + class Summary < OpenAI::Internal::Type::BaseModel # A short summary of the reasoning used by the model when generating the response. sig { returns(String) } attr_accessor :text @@ -73,7 +73,7 @@ module OpenAI # The status of the item. One of `in_progress`, `completed`, or `incomplete`. # Populated when items are returned via API. module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseReasoningItem::Status) } OrSymbol = diff --git a/rbi/lib/openai/models/responses/response_refusal_delta_event.rbi b/rbi/lib/openai/models/responses/response_refusal_delta_event.rbi index cbbd66ed..1fffadb6 100644 --- a/rbi/lib/openai/models/responses/response_refusal_delta_event.rbi +++ b/rbi/lib/openai/models/responses/response_refusal_delta_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseRefusalDeltaEvent < OpenAI::BaseModel + class ResponseRefusalDeltaEvent < OpenAI::Internal::Type::BaseModel # The index of the content part that the refusal text is added to. sig { returns(Integer) } attr_accessor :content_index diff --git a/rbi/lib/openai/models/responses/response_refusal_done_event.rbi b/rbi/lib/openai/models/responses/response_refusal_done_event.rbi index 1f24f6df..6b9582e9 100644 --- a/rbi/lib/openai/models/responses/response_refusal_done_event.rbi +++ b/rbi/lib/openai/models/responses/response_refusal_done_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseRefusalDoneEvent < OpenAI::BaseModel + class ResponseRefusalDoneEvent < OpenAI::Internal::Type::BaseModel # The index of the content part that the refusal text is finalized. sig { returns(Integer) } attr_accessor :content_index diff --git a/rbi/lib/openai/models/responses/response_retrieve_params.rbi b/rbi/lib/openai/models/responses/response_retrieve_params.rbi index a6c61a0f..d1308648 100644 --- a/rbi/lib/openai/models/responses/response_retrieve_params.rbi +++ b/rbi/lib/openai/models/responses/response_retrieve_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseRetrieveParams < OpenAI::BaseModel + class ResponseRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -18,7 +18,7 @@ module OpenAI sig do params( include: T::Array[OpenAI::Models::Responses::ResponseIncludable::OrSymbol], - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end diff --git a/rbi/lib/openai/models/responses/response_status.rbi b/rbi/lib/openai/models/responses/response_status.rbi index 138c0188..c69a253e 100644 --- a/rbi/lib/openai/models/responses/response_status.rbi +++ b/rbi/lib/openai/models/responses/response_status.rbi @@ -6,7 +6,7 @@ module OpenAI # The status of the response generation. One of `completed`, `failed`, # `in_progress`, or `incomplete`. module ResponseStatus - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ResponseStatus) } OrSymbol = diff --git a/rbi/lib/openai/models/responses/response_stream_event.rbi b/rbi/lib/openai/models/responses/response_stream_event.rbi index e04029c8..ced70ddc 100644 --- a/rbi/lib/openai/models/responses/response_stream_event.rbi +++ b/rbi/lib/openai/models/responses/response_stream_event.rbi @@ -5,7 +5,7 @@ module OpenAI module Responses # Emitted when there is a partial audio response. module ResponseStreamEvent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/responses/response_text_annotation_delta_event.rbi b/rbi/lib/openai/models/responses/response_text_annotation_delta_event.rbi index 689a3028..e448a3e2 100644 --- a/rbi/lib/openai/models/responses/response_text_annotation_delta_event.rbi +++ b/rbi/lib/openai/models/responses/response_text_annotation_delta_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseTextAnnotationDeltaEvent < OpenAI::BaseModel + class ResponseTextAnnotationDeltaEvent < OpenAI::Internal::Type::BaseModel # A citation to a file. sig do returns( @@ -41,7 +41,7 @@ module OpenAI params( annotation: T.any( OpenAI::Models::Responses::ResponseTextAnnotationDeltaEvent::Annotation::FileCitation, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ResponseTextAnnotationDeltaEvent::Annotation::URLCitation, OpenAI::Models::Responses::ResponseTextAnnotationDeltaEvent::Annotation::FilePath ), @@ -85,9 +85,9 @@ module OpenAI # A citation to a file. module Annotation - extend OpenAI::Union + extend OpenAI::Internal::Type::Union - class FileCitation < OpenAI::BaseModel + class FileCitation < OpenAI::Internal::Type::BaseModel # The ID of the file. sig { returns(String) } attr_accessor :file_id @@ -110,7 +110,7 @@ module OpenAI end end - class URLCitation < OpenAI::BaseModel + class URLCitation < OpenAI::Internal::Type::BaseModel # The index of the last character of the URL citation in the message. sig { returns(Integer) } attr_accessor :end_index @@ -154,7 +154,7 @@ module OpenAI end end - class FilePath < OpenAI::BaseModel + class FilePath < OpenAI::Internal::Type::BaseModel # The ID of the file. sig { returns(String) } attr_accessor :file_id diff --git a/rbi/lib/openai/models/responses/response_text_config.rbi b/rbi/lib/openai/models/responses/response_text_config.rbi index 1aad6df7..b5d1a866 100644 --- a/rbi/lib/openai/models/responses/response_text_config.rbi +++ b/rbi/lib/openai/models/responses/response_text_config.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseTextConfig < OpenAI::BaseModel + class ResponseTextConfig < OpenAI::Internal::Type::BaseModel # An object specifying the format that the model must output. # # Configuring `{ "type": "json_schema" }` enables Structured Outputs, which @@ -34,7 +34,7 @@ module OpenAI params( format_: T.any( OpenAI::Models::ResponseFormatText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ResponseFormatTextJSONSchemaConfig, OpenAI::Models::ResponseFormatJSONObject ) @@ -52,7 +52,7 @@ module OpenAI params( format_: T.any( OpenAI::Models::ResponseFormatText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ResponseFormatTextJSONSchemaConfig, OpenAI::Models::ResponseFormatJSONObject ) diff --git a/rbi/lib/openai/models/responses/response_text_delta_event.rbi b/rbi/lib/openai/models/responses/response_text_delta_event.rbi index efb4c979..7bbc55cd 100644 --- a/rbi/lib/openai/models/responses/response_text_delta_event.rbi +++ b/rbi/lib/openai/models/responses/response_text_delta_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseTextDeltaEvent < OpenAI::BaseModel + class ResponseTextDeltaEvent < OpenAI::Internal::Type::BaseModel # The index of the content part that the text delta was added to. sig { returns(Integer) } attr_accessor :content_index diff --git a/rbi/lib/openai/models/responses/response_text_done_event.rbi b/rbi/lib/openai/models/responses/response_text_done_event.rbi index b8e7bf9a..743e9572 100644 --- a/rbi/lib/openai/models/responses/response_text_done_event.rbi +++ b/rbi/lib/openai/models/responses/response_text_done_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseTextDoneEvent < OpenAI::BaseModel + class ResponseTextDoneEvent < OpenAI::Internal::Type::BaseModel # The index of the content part that the text content is finalized. sig { returns(Integer) } attr_accessor :content_index diff --git a/rbi/lib/openai/models/responses/response_usage.rbi b/rbi/lib/openai/models/responses/response_usage.rbi index c9ac5ea8..47ce8244 100644 --- a/rbi/lib/openai/models/responses/response_usage.rbi +++ b/rbi/lib/openai/models/responses/response_usage.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseUsage < OpenAI::BaseModel + class ResponseUsage < OpenAI::Internal::Type::BaseModel # The number of input tokens. sig { returns(Integer) } attr_accessor :input_tokens @@ -14,7 +14,7 @@ module OpenAI sig do params( - input_tokens_details: T.any(OpenAI::Models::Responses::ResponseUsage::InputTokensDetails, OpenAI::Internal::Util::AnyHash) + input_tokens_details: T.any(OpenAI::Models::Responses::ResponseUsage::InputTokensDetails, OpenAI::Internal::AnyHash) ) .void end @@ -30,7 +30,7 @@ module OpenAI sig do params( - output_tokens_details: T.any(OpenAI::Models::Responses::ResponseUsage::OutputTokensDetails, OpenAI::Internal::Util::AnyHash) + output_tokens_details: T.any(OpenAI::Models::Responses::ResponseUsage::OutputTokensDetails, OpenAI::Internal::AnyHash) ) .void end @@ -45,9 +45,9 @@ module OpenAI sig do params( input_tokens: Integer, - input_tokens_details: T.any(OpenAI::Models::Responses::ResponseUsage::InputTokensDetails, OpenAI::Internal::Util::AnyHash), + input_tokens_details: T.any(OpenAI::Models::Responses::ResponseUsage::InputTokensDetails, OpenAI::Internal::AnyHash), output_tokens: Integer, - output_tokens_details: T.any(OpenAI::Models::Responses::ResponseUsage::OutputTokensDetails, OpenAI::Internal::Util::AnyHash), + output_tokens_details: T.any(OpenAI::Models::Responses::ResponseUsage::OutputTokensDetails, OpenAI::Internal::AnyHash), total_tokens: Integer ) .returns(T.attached_class) @@ -76,7 +76,7 @@ module OpenAI def to_hash end - class InputTokensDetails < OpenAI::BaseModel + class InputTokensDetails < OpenAI::Internal::Type::BaseModel # The number of tokens that were retrieved from the cache. # [More on prompt caching](https://platform.openai.com/docs/guides/prompt-caching). sig { returns(Integer) } @@ -92,7 +92,7 @@ module OpenAI end end - class OutputTokensDetails < OpenAI::BaseModel + class OutputTokensDetails < OpenAI::Internal::Type::BaseModel # The number of reasoning tokens. sig { returns(Integer) } attr_accessor :reasoning_tokens diff --git a/rbi/lib/openai/models/responses/response_web_search_call_completed_event.rbi b/rbi/lib/openai/models/responses/response_web_search_call_completed_event.rbi index b02a5b1f..160537d1 100644 --- a/rbi/lib/openai/models/responses/response_web_search_call_completed_event.rbi +++ b/rbi/lib/openai/models/responses/response_web_search_call_completed_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseWebSearchCallCompletedEvent < OpenAI::BaseModel + class ResponseWebSearchCallCompletedEvent < OpenAI::Internal::Type::BaseModel # Unique ID for the output item associated with the web search call. sig { returns(String) } attr_accessor :item_id diff --git a/rbi/lib/openai/models/responses/response_web_search_call_in_progress_event.rbi b/rbi/lib/openai/models/responses/response_web_search_call_in_progress_event.rbi index 11aa3340..428a1aa1 100644 --- a/rbi/lib/openai/models/responses/response_web_search_call_in_progress_event.rbi +++ b/rbi/lib/openai/models/responses/response_web_search_call_in_progress_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseWebSearchCallInProgressEvent < OpenAI::BaseModel + class ResponseWebSearchCallInProgressEvent < OpenAI::Internal::Type::BaseModel # Unique ID for the output item associated with the web search call. sig { returns(String) } attr_accessor :item_id diff --git a/rbi/lib/openai/models/responses/response_web_search_call_searching_event.rbi b/rbi/lib/openai/models/responses/response_web_search_call_searching_event.rbi index c97724d2..6edd9004 100644 --- a/rbi/lib/openai/models/responses/response_web_search_call_searching_event.rbi +++ b/rbi/lib/openai/models/responses/response_web_search_call_searching_event.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ResponseWebSearchCallSearchingEvent < OpenAI::BaseModel + class ResponseWebSearchCallSearchingEvent < OpenAI::Internal::Type::BaseModel # Unique ID for the output item associated with the web search call. sig { returns(String) } attr_accessor :item_id diff --git a/rbi/lib/openai/models/responses/tool.rbi b/rbi/lib/openai/models/responses/tool.rbi index 0f080b1b..edc55ed0 100644 --- a/rbi/lib/openai/models/responses/tool.rbi +++ b/rbi/lib/openai/models/responses/tool.rbi @@ -7,7 +7,7 @@ module OpenAI # the # [file search tool](https://platform.openai.com/docs/guides/tools-file-search). module Tool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig do override diff --git a/rbi/lib/openai/models/responses/tool_choice_function.rbi b/rbi/lib/openai/models/responses/tool_choice_function.rbi index 29aa1b6c..d22afe4e 100644 --- a/rbi/lib/openai/models/responses/tool_choice_function.rbi +++ b/rbi/lib/openai/models/responses/tool_choice_function.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ToolChoiceFunction < OpenAI::BaseModel + class ToolChoiceFunction < OpenAI::Internal::Type::BaseModel # The name of the function to call. sig { returns(String) } attr_accessor :name diff --git a/rbi/lib/openai/models/responses/tool_choice_options.rbi b/rbi/lib/openai/models/responses/tool_choice_options.rbi index 722831af..9eeb2f91 100644 --- a/rbi/lib/openai/models/responses/tool_choice_options.rbi +++ b/rbi/lib/openai/models/responses/tool_choice_options.rbi @@ -12,7 +12,7 @@ module OpenAI # # `required` means the model must call one or more tools. module ToolChoiceOptions - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ToolChoiceOptions) } OrSymbol = diff --git a/rbi/lib/openai/models/responses/tool_choice_types.rbi b/rbi/lib/openai/models/responses/tool_choice_types.rbi index 75251556..9969a8c5 100644 --- a/rbi/lib/openai/models/responses/tool_choice_types.rbi +++ b/rbi/lib/openai/models/responses/tool_choice_types.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class ToolChoiceTypes < OpenAI::BaseModel + class ToolChoiceTypes < OpenAI::Internal::Type::BaseModel # The type of hosted tool the model should to use. Learn more about # [built-in tools](https://platform.openai.com/docs/guides/tools). # @@ -34,7 +34,7 @@ module OpenAI # - `web_search_preview` # - `computer_use_preview` module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::ToolChoiceTypes::Type) } OrSymbol = diff --git a/rbi/lib/openai/models/responses/web_search_tool.rbi b/rbi/lib/openai/models/responses/web_search_tool.rbi index ba497fb4..5558e18f 100644 --- a/rbi/lib/openai/models/responses/web_search_tool.rbi +++ b/rbi/lib/openai/models/responses/web_search_tool.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Responses - class WebSearchTool < OpenAI::BaseModel + class WebSearchTool < OpenAI::Internal::Type::BaseModel # The type of the web search tool. One of: # # - `web_search_preview` @@ -24,7 +24,7 @@ module OpenAI sig do params( - user_location: T.nilable(T.any(OpenAI::Models::Responses::WebSearchTool::UserLocation, OpenAI::Internal::Util::AnyHash)) + user_location: T.nilable(T.any(OpenAI::Models::Responses::WebSearchTool::UserLocation, OpenAI::Internal::AnyHash)) ) .void end @@ -37,7 +37,7 @@ module OpenAI params( type: OpenAI::Models::Responses::WebSearchTool::Type::OrSymbol, search_context_size: OpenAI::Models::Responses::WebSearchTool::SearchContextSize::OrSymbol, - user_location: T.nilable(T.any(OpenAI::Models::Responses::WebSearchTool::UserLocation, OpenAI::Internal::Util::AnyHash)) + user_location: T.nilable(T.any(OpenAI::Models::Responses::WebSearchTool::UserLocation, OpenAI::Internal::AnyHash)) ) .returns(T.attached_class) end @@ -62,7 +62,7 @@ module OpenAI # - `web_search_preview` # - `web_search_preview_2025_03_11` module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::WebSearchTool::Type) } OrSymbol = @@ -81,7 +81,7 @@ module OpenAI # High level guidance for the amount of context window space to use for the # search. One of `low`, `medium`, or `high`. `medium` is the default. module SearchContextSize - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Responses::WebSearchTool::SearchContextSize) } @@ -97,7 +97,7 @@ module OpenAI end end - class UserLocation < OpenAI::BaseModel + class UserLocation < OpenAI::Internal::Type::BaseModel # The type of location approximation. Always `approximate`. sig { returns(Symbol) } attr_accessor :type diff --git a/rbi/lib/openai/models/responses_model.rbi b/rbi/lib/openai/models/responses_model.rbi index 0c8b5a11..841c3432 100644 --- a/rbi/lib/openai/models/responses_model.rbi +++ b/rbi/lib/openai/models/responses_model.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module ResponsesModel - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, OpenAI::Models::ChatModel::OrSymbol, OpenAI::Models::ResponsesModel::OrSymbol]) } def self.variants diff --git a/rbi/lib/openai/models/static_file_chunking_strategy.rbi b/rbi/lib/openai/models/static_file_chunking_strategy.rbi index bb5abe4d..8ac37465 100644 --- a/rbi/lib/openai/models/static_file_chunking_strategy.rbi +++ b/rbi/lib/openai/models/static_file_chunking_strategy.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class StaticFileChunkingStrategy < OpenAI::BaseModel + class StaticFileChunkingStrategy < OpenAI::Internal::Type::BaseModel # The number of tokens that overlap between chunks. The default value is `400`. # # Note that the overlap must not exceed half of `max_chunk_size_tokens`. diff --git a/rbi/lib/openai/models/static_file_chunking_strategy_object.rbi b/rbi/lib/openai/models/static_file_chunking_strategy_object.rbi index 8525371f..4976ca0e 100644 --- a/rbi/lib/openai/models/static_file_chunking_strategy_object.rbi +++ b/rbi/lib/openai/models/static_file_chunking_strategy_object.rbi @@ -2,11 +2,11 @@ module OpenAI module Models - class StaticFileChunkingStrategyObject < OpenAI::BaseModel + class StaticFileChunkingStrategyObject < OpenAI::Internal::Type::BaseModel sig { returns(OpenAI::Models::StaticFileChunkingStrategy) } attr_reader :static - sig { params(static: T.any(OpenAI::Models::StaticFileChunkingStrategy, OpenAI::Internal::Util::AnyHash)).void } + sig { params(static: T.any(OpenAI::Models::StaticFileChunkingStrategy, OpenAI::Internal::AnyHash)).void } attr_writer :static # Always `static`. @@ -14,10 +14,7 @@ module OpenAI attr_accessor :type sig do - params( - static: T.any(OpenAI::Models::StaticFileChunkingStrategy, OpenAI::Internal::Util::AnyHash), - type: Symbol - ) + params(static: T.any(OpenAI::Models::StaticFileChunkingStrategy, OpenAI::Internal::AnyHash), type: Symbol) .returns(T.attached_class) end def self.new(static:, type: :static) diff --git a/rbi/lib/openai/models/static_file_chunking_strategy_object_param.rbi b/rbi/lib/openai/models/static_file_chunking_strategy_object_param.rbi index fe747e98..b3f1b26c 100644 --- a/rbi/lib/openai/models/static_file_chunking_strategy_object_param.rbi +++ b/rbi/lib/openai/models/static_file_chunking_strategy_object_param.rbi @@ -2,11 +2,11 @@ module OpenAI module Models - class StaticFileChunkingStrategyObjectParam < OpenAI::BaseModel + class StaticFileChunkingStrategyObjectParam < OpenAI::Internal::Type::BaseModel sig { returns(OpenAI::Models::StaticFileChunkingStrategy) } attr_reader :static - sig { params(static: T.any(OpenAI::Models::StaticFileChunkingStrategy, OpenAI::Internal::Util::AnyHash)).void } + sig { params(static: T.any(OpenAI::Models::StaticFileChunkingStrategy, OpenAI::Internal::AnyHash)).void } attr_writer :static # Always `static`. @@ -15,10 +15,7 @@ module OpenAI # Customize your own chunking strategy by setting chunk size and chunk overlap. sig do - params( - static: T.any(OpenAI::Models::StaticFileChunkingStrategy, OpenAI::Internal::Util::AnyHash), - type: Symbol - ) + params(static: T.any(OpenAI::Models::StaticFileChunkingStrategy, OpenAI::Internal::AnyHash), type: Symbol) .returns(T.attached_class) end def self.new(static:, type: :static) diff --git a/rbi/lib/openai/models/upload.rbi b/rbi/lib/openai/models/upload.rbi index feefd876..e3342969 100644 --- a/rbi/lib/openai/models/upload.rbi +++ b/rbi/lib/openai/models/upload.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class Upload < OpenAI::BaseModel + class Upload < OpenAI::Internal::Type::BaseModel # The Upload unique identifier, which can be referenced in API endpoints. sig { returns(String) } attr_accessor :id @@ -41,7 +41,7 @@ module OpenAI sig { returns(T.nilable(OpenAI::Models::FileObject)) } attr_reader :file - sig { params(file: T.nilable(T.any(OpenAI::Models::FileObject, OpenAI::Internal::Util::AnyHash))).void } + sig { params(file: T.nilable(T.any(OpenAI::Models::FileObject, OpenAI::Internal::AnyHash))).void } attr_writer :file # The Upload object can accept byte chunks in the form of Parts. @@ -54,7 +54,7 @@ module OpenAI filename: String, purpose: String, status: OpenAI::Models::Upload::Status::OrSymbol, - file: T.nilable(T.any(OpenAI::Models::FileObject, OpenAI::Internal::Util::AnyHash)), + file: T.nilable(T.any(OpenAI::Models::FileObject, OpenAI::Internal::AnyHash)), object: Symbol ) .returns(T.attached_class) @@ -93,7 +93,7 @@ module OpenAI # The status of the Upload. module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::Upload::Status) } OrSymbol = T.type_alias { T.any(Symbol, String, OpenAI::Models::Upload::Status::TaggedSymbol) } diff --git a/rbi/lib/openai/models/upload_cancel_params.rbi b/rbi/lib/openai/models/upload_cancel_params.rbi index 706ec1dd..f58f71bc 100644 --- a/rbi/lib/openai/models/upload_cancel_params.rbi +++ b/rbi/lib/openai/models/upload_cancel_params.rbi @@ -2,13 +2,17 @@ module OpenAI module Models - class UploadCancelParams < OpenAI::BaseModel + class UploadCancelParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters sig do - params(request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) - .returns(T.attached_class) + params( + request_options: T.any( + OpenAI::RequestOptions, + OpenAI::Internal::AnyHash + ) + ).returns(T.attached_class) end def self.new(request_options: {}) end diff --git a/rbi/lib/openai/models/upload_complete_params.rbi b/rbi/lib/openai/models/upload_complete_params.rbi index 9558d974..3478dd28 100644 --- a/rbi/lib/openai/models/upload_complete_params.rbi +++ b/rbi/lib/openai/models/upload_complete_params.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class UploadCompleteParams < OpenAI::BaseModel + class UploadCompleteParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -22,7 +22,7 @@ module OpenAI params( part_ids: T::Array[String], md5: String, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end diff --git a/rbi/lib/openai/models/upload_create_params.rbi b/rbi/lib/openai/models/upload_create_params.rbi index 50f627d5..aacc2ab0 100644 --- a/rbi/lib/openai/models/upload_create_params.rbi +++ b/rbi/lib/openai/models/upload_create_params.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class UploadCreateParams < OpenAI::BaseModel + class UploadCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -34,7 +34,7 @@ module OpenAI filename: String, mime_type: String, purpose: OpenAI::Models::FilePurpose::OrSymbol, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end diff --git a/rbi/lib/openai/models/uploads/part_create_params.rbi b/rbi/lib/openai/models/uploads/part_create_params.rbi index b840d72b..43f97062 100644 --- a/rbi/lib/openai/models/uploads/part_create_params.rbi +++ b/rbi/lib/openai/models/uploads/part_create_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Uploads - class PartCreateParams < OpenAI::BaseModel + class PartCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -14,7 +14,7 @@ module OpenAI sig do params( data: T.any(IO, StringIO), - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end diff --git a/rbi/lib/openai/models/uploads/upload_part.rbi b/rbi/lib/openai/models/uploads/upload_part.rbi index 43223349..5e5b8819 100644 --- a/rbi/lib/openai/models/uploads/upload_part.rbi +++ b/rbi/lib/openai/models/uploads/upload_part.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module Uploads - class UploadPart < OpenAI::BaseModel + class UploadPart < OpenAI::Internal::Type::BaseModel # The upload Part unique identifier, which can be referenced in API endpoints. sig { returns(String) } attr_accessor :id diff --git a/rbi/lib/openai/models/vector_store.rbi b/rbi/lib/openai/models/vector_store.rbi index e0277e39..0c48914f 100644 --- a/rbi/lib/openai/models/vector_store.rbi +++ b/rbi/lib/openai/models/vector_store.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class VectorStore < OpenAI::BaseModel + class VectorStore < OpenAI::Internal::Type::BaseModel # The identifier, which can be referenced in API endpoints. sig { returns(String) } attr_accessor :id @@ -14,7 +14,7 @@ module OpenAI sig { returns(OpenAI::Models::VectorStore::FileCounts) } attr_reader :file_counts - sig { params(file_counts: T.any(OpenAI::Models::VectorStore::FileCounts, OpenAI::Internal::Util::AnyHash)).void } + sig { params(file_counts: T.any(OpenAI::Models::VectorStore::FileCounts, OpenAI::Internal::AnyHash)).void } attr_writer :file_counts # The Unix timestamp (in seconds) for when the vector store was last active. @@ -52,10 +52,7 @@ module OpenAI sig { returns(T.nilable(OpenAI::Models::VectorStore::ExpiresAfter)) } attr_reader :expires_after - sig do - params(expires_after: T.any(OpenAI::Models::VectorStore::ExpiresAfter, OpenAI::Internal::Util::AnyHash)) - .void - end + sig { params(expires_after: T.any(OpenAI::Models::VectorStore::ExpiresAfter, OpenAI::Internal::AnyHash)).void } attr_writer :expires_after # The Unix timestamp (in seconds) for when the vector store will expire. @@ -68,13 +65,13 @@ module OpenAI params( id: String, created_at: Integer, - file_counts: T.any(OpenAI::Models::VectorStore::FileCounts, OpenAI::Internal::Util::AnyHash), + file_counts: T.any(OpenAI::Models::VectorStore::FileCounts, OpenAI::Internal::AnyHash), last_active_at: T.nilable(Integer), metadata: T.nilable(T::Hash[Symbol, String]), name: String, status: OpenAI::Models::VectorStore::Status::OrSymbol, usage_bytes: Integer, - expires_after: T.any(OpenAI::Models::VectorStore::ExpiresAfter, OpenAI::Internal::Util::AnyHash), + expires_after: T.any(OpenAI::Models::VectorStore::ExpiresAfter, OpenAI::Internal::AnyHash), expires_at: T.nilable(Integer), object: Symbol ) @@ -116,7 +113,7 @@ module OpenAI def to_hash end - class FileCounts < OpenAI::BaseModel + class FileCounts < OpenAI::Internal::Type::BaseModel # The number of files that were cancelled. sig { returns(Integer) } attr_accessor :cancelled @@ -168,7 +165,7 @@ module OpenAI # `completed`. A status of `completed` indicates that the vector store is ready # for use. module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::VectorStore::Status) } OrSymbol = T.type_alias { T.any(Symbol, String, OpenAI::Models::VectorStore::Status::TaggedSymbol) } @@ -182,7 +179,7 @@ module OpenAI end end - class ExpiresAfter < OpenAI::BaseModel + class ExpiresAfter < OpenAI::Internal::Type::BaseModel # Anchor timestamp after which the expiration policy applies. Supported anchors: # `last_active_at`. sig { returns(Symbol) } diff --git a/rbi/lib/openai/models/vector_store_create_params.rbi b/rbi/lib/openai/models/vector_store_create_params.rbi index 86965352..36bee1e6 100644 --- a/rbi/lib/openai/models/vector_store_create_params.rbi +++ b/rbi/lib/openai/models/vector_store_create_params.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class VectorStoreCreateParams < OpenAI::BaseModel + class VectorStoreCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -24,7 +24,7 @@ module OpenAI params( chunking_strategy: T.any( OpenAI::Models::AutoFileChunkingStrategyParam, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::StaticFileChunkingStrategyObjectParam ) ) @@ -38,7 +38,7 @@ module OpenAI sig do params( - expires_after: T.any(OpenAI::Models::VectorStoreCreateParams::ExpiresAfter, OpenAI::Internal::Util::AnyHash) + expires_after: T.any(OpenAI::Models::VectorStoreCreateParams::ExpiresAfter, OpenAI::Internal::AnyHash) ) .void end @@ -73,14 +73,14 @@ module OpenAI params( chunking_strategy: T.any( OpenAI::Models::AutoFileChunkingStrategyParam, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::StaticFileChunkingStrategyObjectParam ), - expires_after: T.any(OpenAI::Models::VectorStoreCreateParams::ExpiresAfter, OpenAI::Internal::Util::AnyHash), + expires_after: T.any(OpenAI::Models::VectorStoreCreateParams::ExpiresAfter, OpenAI::Internal::AnyHash), file_ids: T::Array[String], metadata: T.nilable(T::Hash[Symbol, String]), name: String, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -113,7 +113,7 @@ module OpenAI def to_hash end - class ExpiresAfter < OpenAI::BaseModel + class ExpiresAfter < OpenAI::Internal::Type::BaseModel # Anchor timestamp after which the expiration policy applies. Supported anchors: # `last_active_at`. sig { returns(Symbol) } diff --git a/rbi/lib/openai/models/vector_store_delete_params.rbi b/rbi/lib/openai/models/vector_store_delete_params.rbi index 9ea9ee73..9f646bad 100644 --- a/rbi/lib/openai/models/vector_store_delete_params.rbi +++ b/rbi/lib/openai/models/vector_store_delete_params.rbi @@ -2,13 +2,17 @@ module OpenAI module Models - class VectorStoreDeleteParams < OpenAI::BaseModel + class VectorStoreDeleteParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters sig do - params(request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) - .returns(T.attached_class) + params( + request_options: T.any( + OpenAI::RequestOptions, + OpenAI::Internal::AnyHash + ) + ).returns(T.attached_class) end def self.new(request_options: {}) end diff --git a/rbi/lib/openai/models/vector_store_deleted.rbi b/rbi/lib/openai/models/vector_store_deleted.rbi index bd0e83c0..7143c3a2 100644 --- a/rbi/lib/openai/models/vector_store_deleted.rbi +++ b/rbi/lib/openai/models/vector_store_deleted.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class VectorStoreDeleted < OpenAI::BaseModel + class VectorStoreDeleted < OpenAI::Internal::Type::BaseModel sig { returns(String) } attr_accessor :id diff --git a/rbi/lib/openai/models/vector_store_list_params.rbi b/rbi/lib/openai/models/vector_store_list_params.rbi index ef5af052..bbe5e54f 100644 --- a/rbi/lib/openai/models/vector_store_list_params.rbi +++ b/rbi/lib/openai/models/vector_store_list_params.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class VectorStoreListParams < OpenAI::BaseModel + class VectorStoreListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -48,7 +48,7 @@ module OpenAI before: String, limit: Integer, order: OpenAI::Models::VectorStoreListParams::Order::OrSymbol, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -73,7 +73,7 @@ module OpenAI # Sort order by the `created_at` timestamp of the objects. `asc` for ascending # order and `desc` for descending order. module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::VectorStoreListParams::Order) } OrSymbol = diff --git a/rbi/lib/openai/models/vector_store_retrieve_params.rbi b/rbi/lib/openai/models/vector_store_retrieve_params.rbi index f73d277f..a8ea5ee3 100644 --- a/rbi/lib/openai/models/vector_store_retrieve_params.rbi +++ b/rbi/lib/openai/models/vector_store_retrieve_params.rbi @@ -2,13 +2,17 @@ module OpenAI module Models - class VectorStoreRetrieveParams < OpenAI::BaseModel + class VectorStoreRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters sig do - params(request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) - .returns(T.attached_class) + params( + request_options: T.any( + OpenAI::RequestOptions, + OpenAI::Internal::AnyHash + ) + ).returns(T.attached_class) end def self.new(request_options: {}) end diff --git a/rbi/lib/openai/models/vector_store_search_params.rbi b/rbi/lib/openai/models/vector_store_search_params.rbi index 15dfc27f..7019fb58 100644 --- a/rbi/lib/openai/models/vector_store_search_params.rbi +++ b/rbi/lib/openai/models/vector_store_search_params.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class VectorStoreSearchParams < OpenAI::BaseModel + class VectorStoreSearchParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -16,7 +16,7 @@ module OpenAI sig do params( - filters: T.any(OpenAI::Models::ComparisonFilter, OpenAI::Internal::Util::AnyHash, OpenAI::Models::CompoundFilter) + filters: T.any(OpenAI::Models::ComparisonFilter, OpenAI::Internal::AnyHash, OpenAI::Models::CompoundFilter) ) .void end @@ -36,7 +36,7 @@ module OpenAI sig do params( - ranking_options: T.any(OpenAI::Models::VectorStoreSearchParams::RankingOptions, OpenAI::Internal::Util::AnyHash) + ranking_options: T.any(OpenAI::Models::VectorStoreSearchParams::RankingOptions, OpenAI::Internal::AnyHash) ) .void end @@ -52,11 +52,11 @@ module OpenAI sig do params( query: T.any(String, T::Array[String]), - filters: T.any(OpenAI::Models::ComparisonFilter, OpenAI::Internal::Util::AnyHash, OpenAI::Models::CompoundFilter), + filters: T.any(OpenAI::Models::ComparisonFilter, OpenAI::Internal::AnyHash, OpenAI::Models::CompoundFilter), max_num_results: Integer, - ranking_options: T.any(OpenAI::Models::VectorStoreSearchParams::RankingOptions, OpenAI::Internal::Util::AnyHash), + ranking_options: T.any(OpenAI::Models::VectorStoreSearchParams::RankingOptions, OpenAI::Internal::AnyHash), rewrite_query: T::Boolean, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -88,25 +88,25 @@ module OpenAI # A query string for a search module Query - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, T::Array[String]]) } def self.variants end - StringArray = T.let(OpenAI::ArrayOf[String], OpenAI::Internal::Type::Converter) + StringArray = T.let(OpenAI::Internal::Type::ArrayOf[String], OpenAI::Internal::Type::Converter) end # A filter to apply based on file attributes. module Filters - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([OpenAI::Models::ComparisonFilter, OpenAI::Models::CompoundFilter]) } def self.variants end end - class RankingOptions < OpenAI::BaseModel + class RankingOptions < OpenAI::Internal::Type::BaseModel sig { returns(T.nilable(OpenAI::Models::VectorStoreSearchParams::RankingOptions::Ranker::OrSymbol)) } attr_reader :ranker @@ -140,7 +140,7 @@ module OpenAI end module Ranker - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::VectorStoreSearchParams::RankingOptions::Ranker) } diff --git a/rbi/lib/openai/models/vector_store_search_response.rbi b/rbi/lib/openai/models/vector_store_search_response.rbi index 05fd7444..a024e03c 100644 --- a/rbi/lib/openai/models/vector_store_search_response.rbi +++ b/rbi/lib/openai/models/vector_store_search_response.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class VectorStoreSearchResponse < OpenAI::BaseModel + class VectorStoreSearchResponse < OpenAI::Internal::Type::BaseModel # Set of 16 key-value pairs that can be attached to an object. This can be useful # for storing additional information about the object in a structured format, and # querying for objects via API or the dashboard. Keys are strings with a maximum @@ -30,7 +30,7 @@ module OpenAI sig do params( attributes: T.nilable(T::Hash[Symbol, T.any(String, Float, T::Boolean)]), - content: T::Array[T.any(OpenAI::Models::VectorStoreSearchResponse::Content, OpenAI::Internal::Util::AnyHash)], + content: T::Array[T.any(OpenAI::Models::VectorStoreSearchResponse::Content, OpenAI::Internal::AnyHash)], file_id: String, filename: String, score: Float @@ -56,14 +56,14 @@ module OpenAI end module Attribute - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, Float, T::Boolean]) } def self.variants end end - class Content < OpenAI::BaseModel + class Content < OpenAI::Internal::Type::BaseModel # The text content returned from search. sig { returns(String) } attr_accessor :text @@ -88,7 +88,7 @@ module OpenAI # The type of content. module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::VectorStoreSearchResponse::Content::Type) } OrSymbol = diff --git a/rbi/lib/openai/models/vector_store_update_params.rbi b/rbi/lib/openai/models/vector_store_update_params.rbi index 1982bdfd..7f3e0224 100644 --- a/rbi/lib/openai/models/vector_store_update_params.rbi +++ b/rbi/lib/openai/models/vector_store_update_params.rbi @@ -2,7 +2,7 @@ module OpenAI module Models - class VectorStoreUpdateParams < OpenAI::BaseModel + class VectorStoreUpdateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -12,7 +12,7 @@ module OpenAI sig do params( - expires_after: T.nilable(T.any(OpenAI::Models::VectorStoreUpdateParams::ExpiresAfter, OpenAI::Internal::Util::AnyHash)) + expires_after: T.nilable(T.any(OpenAI::Models::VectorStoreUpdateParams::ExpiresAfter, OpenAI::Internal::AnyHash)) ) .void end @@ -33,10 +33,10 @@ module OpenAI sig do params( - expires_after: T.nilable(T.any(OpenAI::Models::VectorStoreUpdateParams::ExpiresAfter, OpenAI::Internal::Util::AnyHash)), + expires_after: T.nilable(T.any(OpenAI::Models::VectorStoreUpdateParams::ExpiresAfter, OpenAI::Internal::AnyHash)), metadata: T.nilable(T::Hash[Symbol, String]), name: T.nilable(String), - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -57,7 +57,7 @@ module OpenAI def to_hash end - class ExpiresAfter < OpenAI::BaseModel + class ExpiresAfter < OpenAI::Internal::Type::BaseModel # Anchor timestamp after which the expiration policy applies. Supported anchors: # `last_active_at`. sig { returns(Symbol) } diff --git a/rbi/lib/openai/models/vector_stores/file_batch_cancel_params.rbi b/rbi/lib/openai/models/vector_stores/file_batch_cancel_params.rbi index b8985360..3528e39d 100644 --- a/rbi/lib/openai/models/vector_stores/file_batch_cancel_params.rbi +++ b/rbi/lib/openai/models/vector_stores/file_batch_cancel_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module VectorStores - class FileBatchCancelParams < OpenAI::BaseModel + class FileBatchCancelParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -13,7 +13,7 @@ module OpenAI sig do params( vector_store_id: String, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end diff --git a/rbi/lib/openai/models/vector_stores/file_batch_create_params.rbi b/rbi/lib/openai/models/vector_stores/file_batch_create_params.rbi index 447f5e35..2d21b5fc 100644 --- a/rbi/lib/openai/models/vector_stores/file_batch_create_params.rbi +++ b/rbi/lib/openai/models/vector_stores/file_batch_create_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module VectorStores - class FileBatchCreateParams < OpenAI::BaseModel + class FileBatchCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -39,7 +39,7 @@ module OpenAI params( chunking_strategy: T.any( OpenAI::Models::AutoFileChunkingStrategyParam, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::StaticFileChunkingStrategyObjectParam ) ) @@ -53,10 +53,10 @@ module OpenAI attributes: T.nilable(T::Hash[Symbol, T.any(String, Float, T::Boolean)]), chunking_strategy: T.any( OpenAI::Models::AutoFileChunkingStrategyParam, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::StaticFileChunkingStrategyObjectParam ), - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -81,7 +81,7 @@ module OpenAI end module Attribute - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, Float, T::Boolean]) } def self.variants diff --git a/rbi/lib/openai/models/vector_stores/file_batch_list_files_params.rbi b/rbi/lib/openai/models/vector_stores/file_batch_list_files_params.rbi index 07693f31..242a405d 100644 --- a/rbi/lib/openai/models/vector_stores/file_batch_list_files_params.rbi +++ b/rbi/lib/openai/models/vector_stores/file_batch_list_files_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module VectorStores - class FileBatchListFilesParams < OpenAI::BaseModel + class FileBatchListFilesParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -61,7 +61,7 @@ module OpenAI filter: OpenAI::Models::VectorStores::FileBatchListFilesParams::Filter::OrSymbol, limit: Integer, order: OpenAI::Models::VectorStores::FileBatchListFilesParams::Order::OrSymbol, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -95,7 +95,7 @@ module OpenAI # Filter by file status. One of `in_progress`, `completed`, `failed`, `cancelled`. module Filter - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::VectorStores::FileBatchListFilesParams::Filter) } @@ -118,7 +118,7 @@ module OpenAI # Sort order by the `created_at` timestamp of the objects. `asc` for ascending # order and `desc` for descending order. module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::VectorStores::FileBatchListFilesParams::Order) } diff --git a/rbi/lib/openai/models/vector_stores/file_batch_retrieve_params.rbi b/rbi/lib/openai/models/vector_stores/file_batch_retrieve_params.rbi index 1d9ae1fd..465c7bdb 100644 --- a/rbi/lib/openai/models/vector_stores/file_batch_retrieve_params.rbi +++ b/rbi/lib/openai/models/vector_stores/file_batch_retrieve_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module VectorStores - class FileBatchRetrieveParams < OpenAI::BaseModel + class FileBatchRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -13,7 +13,7 @@ module OpenAI sig do params( vector_store_id: String, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end diff --git a/rbi/lib/openai/models/vector_stores/file_content_params.rbi b/rbi/lib/openai/models/vector_stores/file_content_params.rbi index 3889fd16..98263089 100644 --- a/rbi/lib/openai/models/vector_stores/file_content_params.rbi +++ b/rbi/lib/openai/models/vector_stores/file_content_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module VectorStores - class FileContentParams < OpenAI::BaseModel + class FileContentParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -13,7 +13,7 @@ module OpenAI sig do params( vector_store_id: String, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end diff --git a/rbi/lib/openai/models/vector_stores/file_content_response.rbi b/rbi/lib/openai/models/vector_stores/file_content_response.rbi index 1fad3ed9..9b48e97a 100644 --- a/rbi/lib/openai/models/vector_stores/file_content_response.rbi +++ b/rbi/lib/openai/models/vector_stores/file_content_response.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module VectorStores - class FileContentResponse < OpenAI::BaseModel + class FileContentResponse < OpenAI::Internal::Type::BaseModel # The text content sig { returns(T.nilable(String)) } attr_reader :text diff --git a/rbi/lib/openai/models/vector_stores/file_create_params.rbi b/rbi/lib/openai/models/vector_stores/file_create_params.rbi index 221d7a92..dc5f40b1 100644 --- a/rbi/lib/openai/models/vector_stores/file_create_params.rbi +++ b/rbi/lib/openai/models/vector_stores/file_create_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module VectorStores - class FileCreateParams < OpenAI::BaseModel + class FileCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -39,7 +39,7 @@ module OpenAI params( chunking_strategy: T.any( OpenAI::Models::AutoFileChunkingStrategyParam, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::StaticFileChunkingStrategyObjectParam ) ) @@ -53,10 +53,10 @@ module OpenAI attributes: T.nilable(T::Hash[Symbol, T.any(String, Float, T::Boolean)]), chunking_strategy: T.any( OpenAI::Models::AutoFileChunkingStrategyParam, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::StaticFileChunkingStrategyObjectParam ), - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -81,7 +81,7 @@ module OpenAI end module Attribute - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, Float, T::Boolean]) } def self.variants diff --git a/rbi/lib/openai/models/vector_stores/file_delete_params.rbi b/rbi/lib/openai/models/vector_stores/file_delete_params.rbi index 572a5963..940154a6 100644 --- a/rbi/lib/openai/models/vector_stores/file_delete_params.rbi +++ b/rbi/lib/openai/models/vector_stores/file_delete_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module VectorStores - class FileDeleteParams < OpenAI::BaseModel + class FileDeleteParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -13,7 +13,7 @@ module OpenAI sig do params( vector_store_id: String, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end diff --git a/rbi/lib/openai/models/vector_stores/file_list_params.rbi b/rbi/lib/openai/models/vector_stores/file_list_params.rbi index 9f4a5c71..a13740f0 100644 --- a/rbi/lib/openai/models/vector_stores/file_list_params.rbi +++ b/rbi/lib/openai/models/vector_stores/file_list_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module VectorStores - class FileListParams < OpenAI::BaseModel + class FileListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -57,7 +57,7 @@ module OpenAI filter: OpenAI::Models::VectorStores::FileListParams::Filter::OrSymbol, limit: Integer, order: OpenAI::Models::VectorStores::FileListParams::Order::OrSymbol, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -82,7 +82,7 @@ module OpenAI # Filter by file status. One of `in_progress`, `completed`, `failed`, `cancelled`. module Filter - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::VectorStores::FileListParams::Filter) } OrSymbol = @@ -101,7 +101,7 @@ module OpenAI # Sort order by the `created_at` timestamp of the objects. `asc` for ascending # order and `desc` for descending order. module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::VectorStores::FileListParams::Order) } OrSymbol = diff --git a/rbi/lib/openai/models/vector_stores/file_retrieve_params.rbi b/rbi/lib/openai/models/vector_stores/file_retrieve_params.rbi index 48f7edba..553291b7 100644 --- a/rbi/lib/openai/models/vector_stores/file_retrieve_params.rbi +++ b/rbi/lib/openai/models/vector_stores/file_retrieve_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module VectorStores - class FileRetrieveParams < OpenAI::BaseModel + class FileRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -13,7 +13,7 @@ module OpenAI sig do params( vector_store_id: String, - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end diff --git a/rbi/lib/openai/models/vector_stores/file_update_params.rbi b/rbi/lib/openai/models/vector_stores/file_update_params.rbi index 294449c4..eefd1019 100644 --- a/rbi/lib/openai/models/vector_stores/file_update_params.rbi +++ b/rbi/lib/openai/models/vector_stores/file_update_params.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module VectorStores - class FileUpdateParams < OpenAI::BaseModel + class FileUpdateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -22,7 +22,7 @@ module OpenAI params( vector_store_id: String, attributes: T.nilable(T::Hash[Symbol, T.any(String, Float, T::Boolean)]), - request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash) + request_options: T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash) ) .returns(T.attached_class) end @@ -43,7 +43,7 @@ module OpenAI end module Attribute - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, Float, T::Boolean]) } def self.variants diff --git a/rbi/lib/openai/models/vector_stores/vector_store_file.rbi b/rbi/lib/openai/models/vector_stores/vector_store_file.rbi index 30c4c023..81d9bc52 100644 --- a/rbi/lib/openai/models/vector_stores/vector_store_file.rbi +++ b/rbi/lib/openai/models/vector_stores/vector_store_file.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module VectorStores - class VectorStoreFile < OpenAI::BaseModel + class VectorStoreFile < OpenAI::Internal::Type::BaseModel # The identifier, which can be referenced in API endpoints. sig { returns(String) } attr_accessor :id @@ -19,9 +19,7 @@ module OpenAI sig do params( - last_error: T.nilable( - T.any(OpenAI::Models::VectorStores::VectorStoreFile::LastError, OpenAI::Internal::Util::AnyHash) - ) + last_error: T.nilable(T.any(OpenAI::Models::VectorStores::VectorStoreFile::LastError, OpenAI::Internal::AnyHash)) ) .void end @@ -71,7 +69,7 @@ module OpenAI params( chunking_strategy: T.any( OpenAI::Models::StaticFileChunkingStrategyObject, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::OtherFileChunkingStrategyObject ) ) @@ -84,16 +82,14 @@ module OpenAI params( id: String, created_at: Integer, - last_error: T.nilable( - T.any(OpenAI::Models::VectorStores::VectorStoreFile::LastError, OpenAI::Internal::Util::AnyHash) - ), + last_error: T.nilable(T.any(OpenAI::Models::VectorStores::VectorStoreFile::LastError, OpenAI::Internal::AnyHash)), status: OpenAI::Models::VectorStores::VectorStoreFile::Status::OrSymbol, usage_bytes: Integer, vector_store_id: String, attributes: T.nilable(T::Hash[Symbol, T.any(String, Float, T::Boolean)]), chunking_strategy: T.any( OpenAI::Models::StaticFileChunkingStrategyObject, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::OtherFileChunkingStrategyObject ), object: Symbol @@ -132,7 +128,7 @@ module OpenAI def to_hash end - class LastError < OpenAI::BaseModel + class LastError < OpenAI::Internal::Type::BaseModel # One of `server_error` or `rate_limit_exceeded`. sig { returns(OpenAI::Models::VectorStores::VectorStoreFile::LastError::Code::TaggedSymbol) } attr_accessor :code @@ -161,7 +157,7 @@ module OpenAI # One of `server_error` or `rate_limit_exceeded`. module Code - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::VectorStores::VectorStoreFile::LastError::Code) } @@ -185,7 +181,7 @@ module OpenAI # `completed`, `cancelled`, or `failed`. The status `completed` indicates that the # vector store file is ready for use. module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::VectorStores::VectorStoreFile::Status) } OrSymbol = @@ -202,7 +198,7 @@ module OpenAI end module Attribute - extend OpenAI::Union + extend OpenAI::Internal::Type::Union sig { override.returns([String, Float, T::Boolean]) } def self.variants diff --git a/rbi/lib/openai/models/vector_stores/vector_store_file_batch.rbi b/rbi/lib/openai/models/vector_stores/vector_store_file_batch.rbi index b71f8619..6f9eb44c 100644 --- a/rbi/lib/openai/models/vector_stores/vector_store_file_batch.rbi +++ b/rbi/lib/openai/models/vector_stores/vector_store_file_batch.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module VectorStores - class VectorStoreFileBatch < OpenAI::BaseModel + class VectorStoreFileBatch < OpenAI::Internal::Type::BaseModel # The identifier, which can be referenced in API endpoints. sig { returns(String) } attr_accessor :id @@ -18,7 +18,7 @@ module OpenAI sig do params( - file_counts: T.any(OpenAI::Models::VectorStores::VectorStoreFileBatch::FileCounts, OpenAI::Internal::Util::AnyHash) + file_counts: T.any(OpenAI::Models::VectorStores::VectorStoreFileBatch::FileCounts, OpenAI::Internal::AnyHash) ) .void end @@ -45,7 +45,7 @@ module OpenAI params( id: String, created_at: Integer, - file_counts: T.any(OpenAI::Models::VectorStores::VectorStoreFileBatch::FileCounts, OpenAI::Internal::Util::AnyHash), + file_counts: T.any(OpenAI::Models::VectorStores::VectorStoreFileBatch::FileCounts, OpenAI::Internal::AnyHash), status: OpenAI::Models::VectorStores::VectorStoreFileBatch::Status::OrSymbol, vector_store_id: String, object: Symbol @@ -78,7 +78,7 @@ module OpenAI def to_hash end - class FileCounts < OpenAI::BaseModel + class FileCounts < OpenAI::Internal::Type::BaseModel # The number of files that where cancelled. sig { returns(Integer) } attr_accessor :cancelled @@ -129,7 +129,7 @@ module OpenAI # The status of the vector store files batch, which can be either `in_progress`, # `completed`, `cancelled` or `failed`. module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TaggedSymbol = T.type_alias { T.all(Symbol, OpenAI::Models::VectorStores::VectorStoreFileBatch::Status) } OrSymbol = diff --git a/rbi/lib/openai/models/vector_stores/vector_store_file_deleted.rbi b/rbi/lib/openai/models/vector_stores/vector_store_file_deleted.rbi index 6c4d25fa..702d956c 100644 --- a/rbi/lib/openai/models/vector_stores/vector_store_file_deleted.rbi +++ b/rbi/lib/openai/models/vector_stores/vector_store_file_deleted.rbi @@ -3,7 +3,7 @@ module OpenAI module Models module VectorStores - class VectorStoreFileDeleted < OpenAI::BaseModel + class VectorStoreFileDeleted < OpenAI::Internal::Type::BaseModel sig { returns(String) } attr_accessor :id diff --git a/rbi/lib/openai/request_options.rbi b/rbi/lib/openai/request_options.rbi index 9c665553..017b80bf 100644 --- a/rbi/lib/openai/request_options.rbi +++ b/rbi/lib/openai/request_options.rbi @@ -6,7 +6,7 @@ module OpenAI # # When making a request, you can pass an actual {RequestOptions} instance, or # simply pass a Hash with symbol keys matching the attributes on this class. - class RequestOptions < OpenAI::BaseModel + class RequestOptions < OpenAI::Internal::Type::BaseModel # @api private sig { params(opts: T.any(T.self_type, T::Hash[Symbol, T.anything])).void } def self.validate!(opts) @@ -41,7 +41,7 @@ module OpenAI attr_accessor :timeout # Returns a new instance of RequestOptions. - sig { params(values: OpenAI::Internal::Util::AnyHash).returns(T.attached_class) } + sig { params(values: OpenAI::Internal::AnyHash).returns(T.attached_class) } def self.new(values = {}) end end diff --git a/rbi/lib/openai/resources/audio/speech.rbi b/rbi/lib/openai/resources/audio/speech.rbi index 9b5892ad..43967cf5 100644 --- a/rbi/lib/openai/resources/audio/speech.rbi +++ b/rbi/lib/openai/resources/audio/speech.rbi @@ -13,7 +13,7 @@ module OpenAI instructions: String, response_format: OpenAI::Models::Audio::SpeechCreateParams::ResponseFormat::OrSymbol, speed: Float, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(StringIO) end diff --git a/rbi/lib/openai/resources/audio/transcriptions.rbi b/rbi/lib/openai/resources/audio/transcriptions.rbi index 8ec0dc0f..d0f71924 100644 --- a/rbi/lib/openai/resources/audio/transcriptions.rbi +++ b/rbi/lib/openai/resources/audio/transcriptions.rbi @@ -16,7 +16,7 @@ module OpenAI temperature: Float, timestamp_granularities: T::Array[OpenAI::Models::Audio::TranscriptionCreateParams::TimestampGranularity::OrSymbol], stream: T.noreturn, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(T.any(OpenAI::Models::Audio::Transcription, OpenAI::Models::Audio::TranscriptionVerbose)) end @@ -78,7 +78,7 @@ module OpenAI temperature: Float, timestamp_granularities: T::Array[OpenAI::Models::Audio::TranscriptionCreateParams::TimestampGranularity::OrSymbol], stream: T.noreturn, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns( OpenAI::Internal::Stream[ diff --git a/rbi/lib/openai/resources/audio/translations.rbi b/rbi/lib/openai/resources/audio/translations.rbi index 2e07bd58..448c96b3 100644 --- a/rbi/lib/openai/resources/audio/translations.rbi +++ b/rbi/lib/openai/resources/audio/translations.rbi @@ -12,7 +12,7 @@ module OpenAI prompt: String, response_format: OpenAI::Models::Audio::TranslationCreateParams::ResponseFormat::OrSymbol, temperature: Float, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(T.any(OpenAI::Models::Audio::Translation, OpenAI::Models::Audio::TranslationVerbose)) end diff --git a/rbi/lib/openai/resources/batches.rbi b/rbi/lib/openai/resources/batches.rbi index acaa4e55..f723b248 100644 --- a/rbi/lib/openai/resources/batches.rbi +++ b/rbi/lib/openai/resources/batches.rbi @@ -10,7 +10,7 @@ module OpenAI endpoint: OpenAI::Models::BatchCreateParams::Endpoint::OrSymbol, input_file_id: String, metadata: T.nilable(T::Hash[Symbol, String]), - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Batch) end @@ -48,7 +48,7 @@ module OpenAI sig do params( batch_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Batch) end @@ -64,7 +64,7 @@ module OpenAI params( after: String, limit: Integer, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Internal::CursorPage[OpenAI::Models::Batch]) end @@ -87,7 +87,7 @@ module OpenAI sig do params( batch_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Batch) end diff --git a/rbi/lib/openai/resources/beta/assistants.rbi b/rbi/lib/openai/resources/beta/assistants.rbi index 00296adc..bbbe925b 100644 --- a/rbi/lib/openai/resources/beta/assistants.rbi +++ b/rbi/lib/openai/resources/beta/assistants.rbi @@ -17,25 +17,23 @@ module OpenAI T.any( Symbol, OpenAI::Models::ResponseFormatText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::ResponseFormatJSONObject, OpenAI::Models::ResponseFormatJSONSchema ) ), temperature: T.nilable(Float), - tool_resources: T.nilable( - T.any(OpenAI::Models::Beta::AssistantCreateParams::ToolResources, OpenAI::Internal::Util::AnyHash) - ), + tool_resources: T.nilable(T.any(OpenAI::Models::Beta::AssistantCreateParams::ToolResources, OpenAI::Internal::AnyHash)), tools: T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::FileSearchTool, OpenAI::Models::Beta::FunctionTool ) ], top_p: T.nilable(Float), - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Beta::Assistant) end @@ -115,7 +113,7 @@ module OpenAI sig do params( assistant_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Beta::Assistant) end @@ -140,25 +138,23 @@ module OpenAI T.any( Symbol, OpenAI::Models::ResponseFormatText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::ResponseFormatJSONObject, OpenAI::Models::ResponseFormatJSONSchema ) ), temperature: T.nilable(Float), - tool_resources: T.nilable( - T.any(OpenAI::Models::Beta::AssistantUpdateParams::ToolResources, OpenAI::Internal::Util::AnyHash) - ), + tool_resources: T.nilable(T.any(OpenAI::Models::Beta::AssistantUpdateParams::ToolResources, OpenAI::Internal::AnyHash)), tools: T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::FileSearchTool, OpenAI::Models::Beta::FunctionTool ) ], top_p: T.nilable(Float), - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Beta::Assistant) end @@ -243,7 +239,7 @@ module OpenAI before: String, limit: Integer, order: OpenAI::Models::Beta::AssistantListParams::Order::OrSymbol, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Internal::CursorPage[OpenAI::Models::Beta::Assistant]) end @@ -272,7 +268,7 @@ module OpenAI sig do params( assistant_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Beta::AssistantDeleted) end diff --git a/rbi/lib/openai/resources/beta/threads.rbi b/rbi/lib/openai/resources/beta/threads.rbi index 64816d6d..4a22f258 100644 --- a/rbi/lib/openai/resources/beta/threads.rbi +++ b/rbi/lib/openai/resources/beta/threads.rbi @@ -13,10 +13,10 @@ module OpenAI # Create a thread. sig do params( - messages: T::Array[T.any(OpenAI::Models::Beta::ThreadCreateParams::Message, OpenAI::Internal::Util::AnyHash)], + messages: T::Array[T.any(OpenAI::Models::Beta::ThreadCreateParams::Message, OpenAI::Internal::AnyHash)], metadata: T.nilable(T::Hash[Symbol, String]), - tool_resources: T.nilable(T.any(OpenAI::Models::Beta::ThreadCreateParams::ToolResources, OpenAI::Internal::Util::AnyHash)), - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + tool_resources: T.nilable(T.any(OpenAI::Models::Beta::ThreadCreateParams::ToolResources, OpenAI::Internal::AnyHash)), + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Beta::Thread) end @@ -44,7 +44,7 @@ module OpenAI sig do params( thread_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Beta::Thread) end @@ -60,8 +60,8 @@ module OpenAI params( thread_id: String, metadata: T.nilable(T::Hash[Symbol, String]), - tool_resources: T.nilable(T.any(OpenAI::Models::Beta::ThreadUpdateParams::ToolResources, OpenAI::Internal::Util::AnyHash)), - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + tool_resources: T.nilable(T.any(OpenAI::Models::Beta::ThreadUpdateParams::ToolResources, OpenAI::Internal::AnyHash)), + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Beta::Thread) end @@ -88,7 +88,7 @@ module OpenAI sig do params( thread_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Beta::ThreadDeleted) end @@ -113,28 +113,26 @@ module OpenAI T.any( Symbol, OpenAI::Models::ResponseFormatText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::ResponseFormatJSONObject, OpenAI::Models::ResponseFormatJSONSchema ) ), temperature: T.nilable(Float), - thread: T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread, OpenAI::Internal::Util::AnyHash), + thread: T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread, OpenAI::Internal::AnyHash), tool_choice: T.nilable( T.any( OpenAI::Models::Beta::AssistantToolChoiceOption::Auto::OrSymbol, OpenAI::Models::Beta::AssistantToolChoice, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ), - tool_resources: T.nilable( - T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::ToolResources, OpenAI::Internal::Util::AnyHash) - ), + tool_resources: T.nilable(T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::ToolResources, OpenAI::Internal::AnyHash)), tools: T.nilable( T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::FileSearchTool, OpenAI::Models::Beta::FunctionTool ) @@ -142,10 +140,10 @@ module OpenAI ), top_p: T.nilable(Float), truncation_strategy: T.nilable( - T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::TruncationStrategy, OpenAI::Internal::Util::AnyHash) + T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::TruncationStrategy, OpenAI::Internal::AnyHash) ), stream: T.noreturn, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Beta::Threads::Run) end @@ -259,28 +257,26 @@ module OpenAI T.any( Symbol, OpenAI::Models::ResponseFormatText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::ResponseFormatJSONObject, OpenAI::Models::ResponseFormatJSONSchema ) ), temperature: T.nilable(Float), - thread: T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread, OpenAI::Internal::Util::AnyHash), + thread: T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread, OpenAI::Internal::AnyHash), tool_choice: T.nilable( T.any( OpenAI::Models::Beta::AssistantToolChoiceOption::Auto::OrSymbol, OpenAI::Models::Beta::AssistantToolChoice, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ), - tool_resources: T.nilable( - T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::ToolResources, OpenAI::Internal::Util::AnyHash) - ), + tool_resources: T.nilable(T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::ToolResources, OpenAI::Internal::AnyHash)), tools: T.nilable( T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::FileSearchTool, OpenAI::Models::Beta::FunctionTool ) @@ -288,10 +284,10 @@ module OpenAI ), top_p: T.nilable(Float), truncation_strategy: T.nilable( - T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::TruncationStrategy, OpenAI::Internal::Util::AnyHash) + T.any(OpenAI::Models::Beta::ThreadCreateAndRunParams::TruncationStrategy, OpenAI::Internal::AnyHash) ), stream: T.noreturn, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns( OpenAI::Internal::Stream[ diff --git a/rbi/lib/openai/resources/beta/threads/messages.rbi b/rbi/lib/openai/resources/beta/threads/messages.rbi index a412e38b..8585935d 100644 --- a/rbi/lib/openai/resources/beta/threads/messages.rbi +++ b/rbi/lib/openai/resources/beta/threads/messages.rbi @@ -14,7 +14,7 @@ module OpenAI T::Array[ T.any( OpenAI::Models::Beta::Threads::ImageFileContentBlock, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::Threads::ImageURLContentBlock, OpenAI::Models::Beta::Threads::TextContentBlockParam ) @@ -22,10 +22,10 @@ module OpenAI ), role: OpenAI::Models::Beta::Threads::MessageCreateParams::Role::OrSymbol, attachments: T.nilable( - T::Array[T.any(OpenAI::Models::Beta::Threads::MessageCreateParams::Attachment, OpenAI::Internal::Util::AnyHash)] + T::Array[T.any(OpenAI::Models::Beta::Threads::MessageCreateParams::Attachment, OpenAI::Internal::AnyHash)] ), metadata: T.nilable(T::Hash[Symbol, String]), - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Beta::Threads::Message) end @@ -60,7 +60,7 @@ module OpenAI params( message_id: String, thread_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Beta::Threads::Message) end @@ -80,7 +80,7 @@ module OpenAI message_id: String, thread_id: String, metadata: T.nilable(T::Hash[Symbol, String]), - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Beta::Threads::Message) end @@ -109,7 +109,7 @@ module OpenAI limit: Integer, order: OpenAI::Models::Beta::Threads::MessageListParams::Order::OrSymbol, run_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Internal::CursorPage[OpenAI::Models::Beta::Threads::Message]) end @@ -144,7 +144,7 @@ module OpenAI params( message_id: String, thread_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Beta::Threads::MessageDeleted) end diff --git a/rbi/lib/openai/resources/beta/threads/runs.rbi b/rbi/lib/openai/resources/beta/threads/runs.rbi index 43be9231..a1f66f91 100644 --- a/rbi/lib/openai/resources/beta/threads/runs.rbi +++ b/rbi/lib/openai/resources/beta/threads/runs.rbi @@ -16,7 +16,7 @@ module OpenAI include: T::Array[OpenAI::Models::Beta::Threads::Runs::RunStepInclude::OrSymbol], additional_instructions: T.nilable(String), additional_messages: T.nilable( - T::Array[T.any(OpenAI::Models::Beta::Threads::RunCreateParams::AdditionalMessage, OpenAI::Internal::Util::AnyHash)] + T::Array[T.any(OpenAI::Models::Beta::Threads::RunCreateParams::AdditionalMessage, OpenAI::Internal::AnyHash)] ), instructions: T.nilable(String), max_completion_tokens: T.nilable(Integer), @@ -29,7 +29,7 @@ module OpenAI T.any( Symbol, OpenAI::Models::ResponseFormatText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::ResponseFormatJSONObject, OpenAI::Models::ResponseFormatJSONSchema ) @@ -39,14 +39,14 @@ module OpenAI T.any( OpenAI::Models::Beta::AssistantToolChoiceOption::Auto::OrSymbol, OpenAI::Models::Beta::AssistantToolChoice, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ), tools: T.nilable( T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::FileSearchTool, OpenAI::Models::Beta::FunctionTool ) @@ -54,10 +54,10 @@ module OpenAI ), top_p: T.nilable(Float), truncation_strategy: T.nilable( - T.any(OpenAI::Models::Beta::Threads::RunCreateParams::TruncationStrategy, OpenAI::Internal::Util::AnyHash) + T.any(OpenAI::Models::Beta::Threads::RunCreateParams::TruncationStrategy, OpenAI::Internal::AnyHash) ), stream: T.noreturn, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Beta::Threads::Run) end @@ -185,7 +185,7 @@ module OpenAI include: T::Array[OpenAI::Models::Beta::Threads::Runs::RunStepInclude::OrSymbol], additional_instructions: T.nilable(String), additional_messages: T.nilable( - T::Array[T.any(OpenAI::Models::Beta::Threads::RunCreateParams::AdditionalMessage, OpenAI::Internal::Util::AnyHash)] + T::Array[T.any(OpenAI::Models::Beta::Threads::RunCreateParams::AdditionalMessage, OpenAI::Internal::AnyHash)] ), instructions: T.nilable(String), max_completion_tokens: T.nilable(Integer), @@ -198,7 +198,7 @@ module OpenAI T.any( Symbol, OpenAI::Models::ResponseFormatText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::ResponseFormatJSONObject, OpenAI::Models::ResponseFormatJSONSchema ) @@ -208,14 +208,14 @@ module OpenAI T.any( OpenAI::Models::Beta::AssistantToolChoiceOption::Auto::OrSymbol, OpenAI::Models::Beta::AssistantToolChoice, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ) ), tools: T.nilable( T::Array[ T.any( OpenAI::Models::Beta::CodeInterpreterTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Beta::FileSearchTool, OpenAI::Models::Beta::FunctionTool ) @@ -223,10 +223,10 @@ module OpenAI ), top_p: T.nilable(Float), truncation_strategy: T.nilable( - T.any(OpenAI::Models::Beta::Threads::RunCreateParams::TruncationStrategy, OpenAI::Internal::Util::AnyHash) + T.any(OpenAI::Models::Beta::Threads::RunCreateParams::TruncationStrategy, OpenAI::Internal::AnyHash) ), stream: T.noreturn, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns( OpenAI::Internal::Stream[ @@ -380,7 +380,7 @@ module OpenAI params( run_id: String, thread_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Beta::Threads::Run) end @@ -400,7 +400,7 @@ module OpenAI run_id: String, thread_id: String, metadata: T.nilable(T::Hash[Symbol, String]), - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Beta::Threads::Run) end @@ -429,7 +429,7 @@ module OpenAI before: String, limit: Integer, order: OpenAI::Models::Beta::Threads::RunListParams::Order::OrSymbol, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Internal::CursorPage[OpenAI::Models::Beta::Threads::Run]) end @@ -461,7 +461,7 @@ module OpenAI params( run_id: String, thread_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Beta::Threads::Run) end @@ -482,14 +482,9 @@ module OpenAI params( run_id: String, thread_id: String, - tool_outputs: T::Array[ - T.any( - OpenAI::Models::Beta::Threads::RunSubmitToolOutputsParams::ToolOutput, - OpenAI::Internal::Util::AnyHash - ) - ], + tool_outputs: T::Array[T.any(OpenAI::Models::Beta::Threads::RunSubmitToolOutputsParams::ToolOutput, OpenAI::Internal::AnyHash)], stream: T.noreturn, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Beta::Threads::Run) end @@ -517,14 +512,9 @@ module OpenAI params( run_id: String, thread_id: String, - tool_outputs: T::Array[ - T.any( - OpenAI::Models::Beta::Threads::RunSubmitToolOutputsParams::ToolOutput, - OpenAI::Internal::Util::AnyHash - ) - ], + tool_outputs: T::Array[T.any(OpenAI::Models::Beta::Threads::RunSubmitToolOutputsParams::ToolOutput, OpenAI::Internal::AnyHash)], stream: T.noreturn, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns( OpenAI::Internal::Stream[ diff --git a/rbi/lib/openai/resources/beta/threads/runs/steps.rbi b/rbi/lib/openai/resources/beta/threads/runs/steps.rbi index d035232e..81727780 100644 --- a/rbi/lib/openai/resources/beta/threads/runs/steps.rbi +++ b/rbi/lib/openai/resources/beta/threads/runs/steps.rbi @@ -13,7 +13,7 @@ module OpenAI thread_id: String, run_id: String, include: T::Array[OpenAI::Models::Beta::Threads::Runs::RunStepInclude::OrSymbol], - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Beta::Threads::Runs::RunStep) end @@ -47,7 +47,7 @@ module OpenAI include: T::Array[OpenAI::Models::Beta::Threads::Runs::RunStepInclude::OrSymbol], limit: Integer, order: OpenAI::Models::Beta::Threads::Runs::StepListParams::Order::OrSymbol, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Internal::CursorPage[OpenAI::Models::Beta::Threads::Runs::RunStep]) end diff --git a/rbi/lib/openai/resources/chat/completions.rbi b/rbi/lib/openai/resources/chat/completions.rbi index ebf756a5..7a170eb6 100644 --- a/rbi/lib/openai/resources/chat/completions.rbi +++ b/rbi/lib/openai/resources/chat/completions.rbi @@ -29,7 +29,7 @@ module OpenAI messages: T::Array[ T.any( OpenAI::Models::Chat::ChatCompletionDeveloperMessageParam, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Chat::ChatCompletionSystemMessageParam, OpenAI::Models::Chat::ChatCompletionUserMessageParam, OpenAI::Models::Chat::ChatCompletionAssistantMessageParam, @@ -38,14 +38,14 @@ module OpenAI ) ], model: T.any(String, OpenAI::Models::ChatModel::OrSymbol), - audio: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionAudioParam, OpenAI::Internal::Util::AnyHash)), + audio: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionAudioParam, OpenAI::Internal::AnyHash)), frequency_penalty: T.nilable(Float), function_call: T.any( OpenAI::Models::Chat::CompletionCreateParams::FunctionCall::FunctionCallMode::OrSymbol, OpenAI::Models::Chat::ChatCompletionFunctionCallOption, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ), - functions: T::Array[T.any(OpenAI::Models::Chat::CompletionCreateParams::Function, OpenAI::Internal::Util::AnyHash)], + functions: T::Array[T.any(OpenAI::Models::Chat::CompletionCreateParams::Function, OpenAI::Internal::AnyHash)], logit_bias: T.nilable(T::Hash[Symbol, Integer]), logprobs: T.nilable(T::Boolean), max_completion_tokens: T.nilable(Integer), @@ -54,12 +54,12 @@ module OpenAI modalities: T.nilable(T::Array[OpenAI::Models::Chat::CompletionCreateParams::Modality::OrSymbol]), n: T.nilable(Integer), parallel_tool_calls: T::Boolean, - prediction: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionPredictionContent, OpenAI::Internal::Util::AnyHash)), + prediction: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionPredictionContent, OpenAI::Internal::AnyHash)), presence_penalty: T.nilable(Float), reasoning_effort: T.nilable(OpenAI::Models::ReasoningEffort::OrSymbol), response_format: T.any( OpenAI::Models::ResponseFormatText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::ResponseFormatJSONSchema, OpenAI::Models::ResponseFormatJSONObject ), @@ -67,20 +67,20 @@ module OpenAI service_tier: T.nilable(OpenAI::Models::Chat::CompletionCreateParams::ServiceTier::OrSymbol), stop: T.nilable(T.any(String, T::Array[String])), store: T.nilable(T::Boolean), - stream_options: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionStreamOptions, OpenAI::Internal::Util::AnyHash)), + stream_options: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionStreamOptions, OpenAI::Internal::AnyHash)), temperature: T.nilable(Float), tool_choice: T.any( OpenAI::Models::Chat::ChatCompletionToolChoiceOption::Auto::OrSymbol, OpenAI::Models::Chat::ChatCompletionNamedToolChoice, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ), - tools: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionTool, OpenAI::Internal::Util::AnyHash)], + tools: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionTool, OpenAI::Internal::AnyHash)], top_logprobs: T.nilable(Integer), top_p: T.nilable(Float), user: String, - web_search_options: T.any(OpenAI::Models::Chat::CompletionCreateParams::WebSearchOptions, OpenAI::Internal::Util::AnyHash), + web_search_options: T.any(OpenAI::Models::Chat::CompletionCreateParams::WebSearchOptions, OpenAI::Internal::AnyHash), stream: T.noreturn, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Chat::ChatCompletion) end @@ -297,7 +297,7 @@ module OpenAI messages: T::Array[ T.any( OpenAI::Models::Chat::ChatCompletionDeveloperMessageParam, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Chat::ChatCompletionSystemMessageParam, OpenAI::Models::Chat::ChatCompletionUserMessageParam, OpenAI::Models::Chat::ChatCompletionAssistantMessageParam, @@ -306,14 +306,14 @@ module OpenAI ) ], model: T.any(String, OpenAI::Models::ChatModel::OrSymbol), - audio: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionAudioParam, OpenAI::Internal::Util::AnyHash)), + audio: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionAudioParam, OpenAI::Internal::AnyHash)), frequency_penalty: T.nilable(Float), function_call: T.any( OpenAI::Models::Chat::CompletionCreateParams::FunctionCall::FunctionCallMode::OrSymbol, OpenAI::Models::Chat::ChatCompletionFunctionCallOption, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ), - functions: T::Array[T.any(OpenAI::Models::Chat::CompletionCreateParams::Function, OpenAI::Internal::Util::AnyHash)], + functions: T::Array[T.any(OpenAI::Models::Chat::CompletionCreateParams::Function, OpenAI::Internal::AnyHash)], logit_bias: T.nilable(T::Hash[Symbol, Integer]), logprobs: T.nilable(T::Boolean), max_completion_tokens: T.nilable(Integer), @@ -322,12 +322,12 @@ module OpenAI modalities: T.nilable(T::Array[OpenAI::Models::Chat::CompletionCreateParams::Modality::OrSymbol]), n: T.nilable(Integer), parallel_tool_calls: T::Boolean, - prediction: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionPredictionContent, OpenAI::Internal::Util::AnyHash)), + prediction: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionPredictionContent, OpenAI::Internal::AnyHash)), presence_penalty: T.nilable(Float), reasoning_effort: T.nilable(OpenAI::Models::ReasoningEffort::OrSymbol), response_format: T.any( OpenAI::Models::ResponseFormatText, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::ResponseFormatJSONSchema, OpenAI::Models::ResponseFormatJSONObject ), @@ -335,20 +335,20 @@ module OpenAI service_tier: T.nilable(OpenAI::Models::Chat::CompletionCreateParams::ServiceTier::OrSymbol), stop: T.nilable(T.any(String, T::Array[String])), store: T.nilable(T::Boolean), - stream_options: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionStreamOptions, OpenAI::Internal::Util::AnyHash)), + stream_options: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionStreamOptions, OpenAI::Internal::AnyHash)), temperature: T.nilable(Float), tool_choice: T.any( OpenAI::Models::Chat::ChatCompletionToolChoiceOption::Auto::OrSymbol, OpenAI::Models::Chat::ChatCompletionNamedToolChoice, - OpenAI::Internal::Util::AnyHash + OpenAI::Internal::AnyHash ), - tools: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionTool, OpenAI::Internal::Util::AnyHash)], + tools: T::Array[T.any(OpenAI::Models::Chat::ChatCompletionTool, OpenAI::Internal::AnyHash)], top_logprobs: T.nilable(Integer), top_p: T.nilable(Float), user: String, - web_search_options: T.any(OpenAI::Models::Chat::CompletionCreateParams::WebSearchOptions, OpenAI::Internal::Util::AnyHash), + web_search_options: T.any(OpenAI::Models::Chat::CompletionCreateParams::WebSearchOptions, OpenAI::Internal::AnyHash), stream: T.noreturn, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Internal::Stream[OpenAI::Models::Chat::ChatCompletionChunk]) end @@ -548,7 +548,7 @@ module OpenAI sig do params( completion_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Chat::ChatCompletion) end @@ -566,7 +566,7 @@ module OpenAI params( completion_id: String, metadata: T.nilable(T::Hash[Symbol, String]), - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Chat::ChatCompletion) end @@ -593,7 +593,7 @@ module OpenAI metadata: T.nilable(T::Hash[Symbol, String]), model: String, order: OpenAI::Models::Chat::CompletionListParams::Order::OrSymbol, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Internal::CursorPage[OpenAI::Models::Chat::ChatCompletion]) end @@ -620,7 +620,7 @@ module OpenAI sig do params( completion_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Chat::ChatCompletionDeleted) end diff --git a/rbi/lib/openai/resources/chat/completions/messages.rbi b/rbi/lib/openai/resources/chat/completions/messages.rbi index 447ff6c0..440c076c 100644 --- a/rbi/lib/openai/resources/chat/completions/messages.rbi +++ b/rbi/lib/openai/resources/chat/completions/messages.rbi @@ -13,7 +13,7 @@ module OpenAI after: String, limit: Integer, order: OpenAI::Models::Chat::Completions::MessageListParams::Order::OrSymbol, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Internal::CursorPage[OpenAI::Models::Chat::ChatCompletionStoreMessage]) end diff --git a/rbi/lib/openai/resources/completions.rbi b/rbi/lib/openai/resources/completions.rbi index 804db29b..08f12a97 100644 --- a/rbi/lib/openai/resources/completions.rbi +++ b/rbi/lib/openai/resources/completions.rbi @@ -25,13 +25,13 @@ module OpenAI presence_penalty: T.nilable(Float), seed: T.nilable(Integer), stop: T.nilable(T.any(String, T::Array[String])), - stream_options: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionStreamOptions, OpenAI::Internal::Util::AnyHash)), + stream_options: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionStreamOptions, OpenAI::Internal::AnyHash)), suffix: T.nilable(String), temperature: T.nilable(Float), top_p: T.nilable(Float), user: String, stream: T.noreturn, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Completion) end @@ -168,13 +168,13 @@ module OpenAI presence_penalty: T.nilable(Float), seed: T.nilable(Integer), stop: T.nilable(T.any(String, T::Array[String])), - stream_options: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionStreamOptions, OpenAI::Internal::Util::AnyHash)), + stream_options: T.nilable(T.any(OpenAI::Models::Chat::ChatCompletionStreamOptions, OpenAI::Internal::AnyHash)), suffix: T.nilable(String), temperature: T.nilable(Float), top_p: T.nilable(Float), user: String, stream: T.noreturn, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Internal::Stream[OpenAI::Models::Completion]) end diff --git a/rbi/lib/openai/resources/embeddings.rbi b/rbi/lib/openai/resources/embeddings.rbi index 57d7fdd6..182cc4d0 100644 --- a/rbi/lib/openai/resources/embeddings.rbi +++ b/rbi/lib/openai/resources/embeddings.rbi @@ -11,7 +11,7 @@ module OpenAI dimensions: Integer, encoding_format: OpenAI::Models::EmbeddingCreateParams::EncodingFormat::OrSymbol, user: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::CreateEmbeddingResponse) end diff --git a/rbi/lib/openai/resources/files.rbi b/rbi/lib/openai/resources/files.rbi index 18c23250..322f5e32 100644 --- a/rbi/lib/openai/resources/files.rbi +++ b/rbi/lib/openai/resources/files.rbi @@ -28,7 +28,7 @@ module OpenAI params( file: T.any(IO, StringIO), purpose: OpenAI::Models::FilePurpose::OrSymbol, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::FileObject) end @@ -48,7 +48,7 @@ module OpenAI sig do params( file_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::FileObject) end @@ -66,7 +66,7 @@ module OpenAI limit: Integer, order: OpenAI::Models::FileListParams::Order::OrSymbol, purpose: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Internal::CursorPage[OpenAI::Models::FileObject]) end @@ -92,7 +92,7 @@ module OpenAI sig do params( file_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::FileDeleted) end @@ -107,7 +107,7 @@ module OpenAI sig do params( file_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(StringIO) end diff --git a/rbi/lib/openai/resources/fine_tuning/jobs.rbi b/rbi/lib/openai/resources/fine_tuning/jobs.rbi index 81ceb5d7..2370a56a 100644 --- a/rbi/lib/openai/resources/fine_tuning/jobs.rbi +++ b/rbi/lib/openai/resources/fine_tuning/jobs.rbi @@ -18,16 +18,16 @@ module OpenAI params( model: T.any(String, OpenAI::Models::FineTuning::JobCreateParams::Model::OrSymbol), training_file: String, - hyperparameters: T.any(OpenAI::Models::FineTuning::JobCreateParams::Hyperparameters, OpenAI::Internal::Util::AnyHash), + hyperparameters: T.any(OpenAI::Models::FineTuning::JobCreateParams::Hyperparameters, OpenAI::Internal::AnyHash), integrations: T.nilable( - T::Array[T.any(OpenAI::Models::FineTuning::JobCreateParams::Integration, OpenAI::Internal::Util::AnyHash)] + T::Array[T.any(OpenAI::Models::FineTuning::JobCreateParams::Integration, OpenAI::Internal::AnyHash)] ), metadata: T.nilable(T::Hash[Symbol, String]), - method_: T.any(OpenAI::Models::FineTuning::JobCreateParams::Method, OpenAI::Internal::Util::AnyHash), + method_: T.any(OpenAI::Models::FineTuning::JobCreateParams::Method, OpenAI::Internal::AnyHash), seed: T.nilable(Integer), suffix: T.nilable(String), validation_file: T.nilable(String), - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::FineTuning::FineTuningJob) end @@ -100,7 +100,7 @@ module OpenAI sig do params( fine_tuning_job_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::FineTuning::FineTuningJob) end @@ -117,7 +117,7 @@ module OpenAI after: String, limit: Integer, metadata: T.nilable(T::Hash[Symbol, String]), - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Internal::CursorPage[OpenAI::Models::FineTuning::FineTuningJob]) end @@ -137,7 +137,7 @@ module OpenAI sig do params( fine_tuning_job_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::FineTuning::FineTuningJob) end @@ -154,7 +154,7 @@ module OpenAI fine_tuning_job_id: String, after: String, limit: Integer, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Internal::CursorPage[OpenAI::Models::FineTuning::FineTuningJobEvent]) end diff --git a/rbi/lib/openai/resources/fine_tuning/jobs/checkpoints.rbi b/rbi/lib/openai/resources/fine_tuning/jobs/checkpoints.rbi index 3274da29..d165ca24 100644 --- a/rbi/lib/openai/resources/fine_tuning/jobs/checkpoints.rbi +++ b/rbi/lib/openai/resources/fine_tuning/jobs/checkpoints.rbi @@ -11,7 +11,7 @@ module OpenAI fine_tuning_job_id: String, after: String, limit: Integer, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Internal::CursorPage[OpenAI::Models::FineTuning::Jobs::FineTuningJobCheckpoint]) end diff --git a/rbi/lib/openai/resources/images.rbi b/rbi/lib/openai/resources/images.rbi index 58353329..515ec04c 100644 --- a/rbi/lib/openai/resources/images.rbi +++ b/rbi/lib/openai/resources/images.rbi @@ -12,7 +12,7 @@ module OpenAI response_format: T.nilable(OpenAI::Models::ImageCreateVariationParams::ResponseFormat::OrSymbol), size: T.nilable(OpenAI::Models::ImageCreateVariationParams::Size::OrSymbol), user: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::ImagesResponse) end @@ -52,7 +52,7 @@ module OpenAI response_format: T.nilable(OpenAI::Models::ImageEditParams::ResponseFormat::OrSymbol), size: T.nilable(OpenAI::Models::ImageEditParams::Size::OrSymbol), user: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::ImagesResponse) end @@ -98,7 +98,7 @@ module OpenAI size: T.nilable(OpenAI::Models::ImageGenerateParams::Size::OrSymbol), style: T.nilable(OpenAI::Models::ImageGenerateParams::Style::OrSymbol), user: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::ImagesResponse) end diff --git a/rbi/lib/openai/resources/models.rbi b/rbi/lib/openai/resources/models.rbi index c78bca31..8a21caa0 100644 --- a/rbi/lib/openai/resources/models.rbi +++ b/rbi/lib/openai/resources/models.rbi @@ -8,7 +8,7 @@ module OpenAI sig do params( model: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Model) end @@ -22,7 +22,7 @@ module OpenAI # Lists the currently available models, and provides basic information about each # one such as the owner and availability. sig do - params(request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash))) + params(request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash))) .returns(OpenAI::Internal::Page[OpenAI::Models::Model]) end def list(request_options: {}) @@ -33,7 +33,7 @@ module OpenAI sig do params( model: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::ModelDeleted) end diff --git a/rbi/lib/openai/resources/moderations.rbi b/rbi/lib/openai/resources/moderations.rbi index 1b180936..8b4b2f36 100644 --- a/rbi/lib/openai/resources/moderations.rbi +++ b/rbi/lib/openai/resources/moderations.rbi @@ -13,13 +13,13 @@ module OpenAI T::Array[ T.any( OpenAI::Models::ModerationImageURLInput, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::ModerationTextInput ) ] ), model: T.any(String, OpenAI::Models::ModerationModel::OrSymbol), - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::ModerationCreateResponse) end diff --git a/rbi/lib/openai/resources/responses.rbi b/rbi/lib/openai/resources/responses.rbi index 585d0875..1999f3c5 100644 --- a/rbi/lib/openai/resources/responses.rbi +++ b/rbi/lib/openai/resources/responses.rbi @@ -24,7 +24,7 @@ module OpenAI T::Array[ T.any( OpenAI::Models::Responses::EasyInputMessage, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ResponseInputItem::Message, OpenAI::Models::Responses::ResponseOutputMessage, OpenAI::Models::Responses::ResponseFileSearchToolCall, @@ -45,20 +45,20 @@ module OpenAI metadata: T.nilable(T::Hash[Symbol, String]), parallel_tool_calls: T.nilable(T::Boolean), previous_response_id: T.nilable(String), - reasoning: T.nilable(T.any(OpenAI::Models::Reasoning, OpenAI::Internal::Util::AnyHash)), + reasoning: T.nilable(T.any(OpenAI::Models::Reasoning, OpenAI::Internal::AnyHash)), store: T.nilable(T::Boolean), temperature: T.nilable(Float), - text: T.any(OpenAI::Models::Responses::ResponseTextConfig, OpenAI::Internal::Util::AnyHash), + text: T.any(OpenAI::Models::Responses::ResponseTextConfig, OpenAI::Internal::AnyHash), tool_choice: T.any( OpenAI::Models::Responses::ToolChoiceOptions::OrSymbol, OpenAI::Models::Responses::ToolChoiceTypes, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ToolChoiceFunction ), tools: T::Array[ T.any( OpenAI::Models::Responses::FileSearchTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::FunctionTool, OpenAI::Models::Responses::ComputerTool, OpenAI::Models::Responses::WebSearchTool @@ -68,7 +68,7 @@ module OpenAI truncation: T.nilable(OpenAI::Models::Responses::ResponseCreateParams::Truncation::OrSymbol), user: String, stream: T.noreturn, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Responses::Response) end @@ -202,7 +202,7 @@ module OpenAI T::Array[ T.any( OpenAI::Models::Responses::EasyInputMessage, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ResponseInputItem::Message, OpenAI::Models::Responses::ResponseOutputMessage, OpenAI::Models::Responses::ResponseFileSearchToolCall, @@ -223,20 +223,20 @@ module OpenAI metadata: T.nilable(T::Hash[Symbol, String]), parallel_tool_calls: T.nilable(T::Boolean), previous_response_id: T.nilable(String), - reasoning: T.nilable(T.any(OpenAI::Models::Reasoning, OpenAI::Internal::Util::AnyHash)), + reasoning: T.nilable(T.any(OpenAI::Models::Reasoning, OpenAI::Internal::AnyHash)), store: T.nilable(T::Boolean), temperature: T.nilable(Float), - text: T.any(OpenAI::Models::Responses::ResponseTextConfig, OpenAI::Internal::Util::AnyHash), + text: T.any(OpenAI::Models::Responses::ResponseTextConfig, OpenAI::Internal::AnyHash), tool_choice: T.any( OpenAI::Models::Responses::ToolChoiceOptions::OrSymbol, OpenAI::Models::Responses::ToolChoiceTypes, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::ToolChoiceFunction ), tools: T::Array[ T.any( OpenAI::Models::Responses::FileSearchTool, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::Responses::FunctionTool, OpenAI::Models::Responses::ComputerTool, OpenAI::Models::Responses::WebSearchTool @@ -246,7 +246,7 @@ module OpenAI truncation: T.nilable(OpenAI::Models::Responses::ResponseCreateParams::Truncation::OrSymbol), user: String, stream: T.noreturn, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns( OpenAI::Internal::Stream[ @@ -404,7 +404,7 @@ module OpenAI params( response_id: String, include: T::Array[OpenAI::Models::Responses::ResponseIncludable::OrSymbol], - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Responses::Response) end @@ -422,7 +422,7 @@ module OpenAI sig do params( response_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .void end diff --git a/rbi/lib/openai/resources/responses/input_items.rbi b/rbi/lib/openai/resources/responses/input_items.rbi index 92e7016f..5c6dfbeb 100644 --- a/rbi/lib/openai/resources/responses/input_items.rbi +++ b/rbi/lib/openai/resources/responses/input_items.rbi @@ -13,7 +13,7 @@ module OpenAI include: T::Array[OpenAI::Models::Responses::ResponseIncludable::OrSymbol], limit: Integer, order: OpenAI::Models::Responses::InputItemListParams::Order::OrSymbol, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns( OpenAI::Internal::CursorPage[ diff --git a/rbi/lib/openai/resources/uploads.rbi b/rbi/lib/openai/resources/uploads.rbi index 88ab62ad..ed222226 100644 --- a/rbi/lib/openai/resources/uploads.rbi +++ b/rbi/lib/openai/resources/uploads.rbi @@ -31,7 +31,7 @@ module OpenAI filename: String, mime_type: String, purpose: OpenAI::Models::FilePurpose::OrSymbol, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Upload) end @@ -58,7 +58,7 @@ module OpenAI sig do params( upload_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Upload) end @@ -87,7 +87,7 @@ module OpenAI upload_id: String, part_ids: T::Array[String], md5: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Upload) end diff --git a/rbi/lib/openai/resources/uploads/parts.rbi b/rbi/lib/openai/resources/uploads/parts.rbi index 6d53630a..7aec6363 100644 --- a/rbi/lib/openai/resources/uploads/parts.rbi +++ b/rbi/lib/openai/resources/uploads/parts.rbi @@ -19,7 +19,7 @@ module OpenAI params( upload_id: String, data: T.any(IO, StringIO), - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::Uploads::UploadPart) end diff --git a/rbi/lib/openai/resources/vector_stores.rbi b/rbi/lib/openai/resources/vector_stores.rbi index 74b9b2a3..e64c7552 100644 --- a/rbi/lib/openai/resources/vector_stores.rbi +++ b/rbi/lib/openai/resources/vector_stores.rbi @@ -14,14 +14,14 @@ module OpenAI params( chunking_strategy: T.any( OpenAI::Models::AutoFileChunkingStrategyParam, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::StaticFileChunkingStrategyObjectParam ), - expires_after: T.any(OpenAI::Models::VectorStoreCreateParams::ExpiresAfter, OpenAI::Internal::Util::AnyHash), + expires_after: T.any(OpenAI::Models::VectorStoreCreateParams::ExpiresAfter, OpenAI::Internal::AnyHash), file_ids: T::Array[String], metadata: T.nilable(T::Hash[Symbol, String]), name: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::VectorStore) end @@ -52,7 +52,7 @@ module OpenAI sig do params( vector_store_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::VectorStore) end @@ -67,10 +67,10 @@ module OpenAI sig do params( vector_store_id: String, - expires_after: T.nilable(T.any(OpenAI::Models::VectorStoreUpdateParams::ExpiresAfter, OpenAI::Internal::Util::AnyHash)), + expires_after: T.nilable(T.any(OpenAI::Models::VectorStoreUpdateParams::ExpiresAfter, OpenAI::Internal::AnyHash)), metadata: T.nilable(T::Hash[Symbol, String]), name: T.nilable(String), - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::VectorStore) end @@ -99,7 +99,7 @@ module OpenAI before: String, limit: Integer, order: OpenAI::Models::VectorStoreListParams::Order::OrSymbol, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Internal::CursorPage[OpenAI::Models::VectorStore]) end @@ -128,7 +128,7 @@ module OpenAI sig do params( vector_store_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::VectorStoreDeleted) end @@ -145,11 +145,11 @@ module OpenAI params( vector_store_id: String, query: T.any(String, T::Array[String]), - filters: T.any(OpenAI::Models::ComparisonFilter, OpenAI::Internal::Util::AnyHash, OpenAI::Models::CompoundFilter), + filters: T.any(OpenAI::Models::ComparisonFilter, OpenAI::Internal::AnyHash, OpenAI::Models::CompoundFilter), max_num_results: Integer, - ranking_options: T.any(OpenAI::Models::VectorStoreSearchParams::RankingOptions, OpenAI::Internal::Util::AnyHash), + ranking_options: T.any(OpenAI::Models::VectorStoreSearchParams::RankingOptions, OpenAI::Internal::AnyHash), rewrite_query: T::Boolean, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Internal::Page[OpenAI::Models::VectorStoreSearchResponse]) end diff --git a/rbi/lib/openai/resources/vector_stores/file_batches.rbi b/rbi/lib/openai/resources/vector_stores/file_batches.rbi index 40c66e29..5bb347bf 100644 --- a/rbi/lib/openai/resources/vector_stores/file_batches.rbi +++ b/rbi/lib/openai/resources/vector_stores/file_batches.rbi @@ -12,10 +12,10 @@ module OpenAI attributes: T.nilable(T::Hash[Symbol, T.any(String, Float, T::Boolean)]), chunking_strategy: T.any( OpenAI::Models::AutoFileChunkingStrategyParam, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::StaticFileChunkingStrategyObjectParam ), - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::VectorStores::VectorStoreFileBatch) end @@ -44,7 +44,7 @@ module OpenAI params( batch_id: String, vector_store_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::VectorStores::VectorStoreFileBatch) end @@ -63,7 +63,7 @@ module OpenAI params( batch_id: String, vector_store_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::VectorStores::VectorStoreFileBatch) end @@ -86,7 +86,7 @@ module OpenAI filter: OpenAI::Models::VectorStores::FileBatchListFilesParams::Filter::OrSymbol, limit: Integer, order: OpenAI::Models::VectorStores::FileBatchListFilesParams::Order::OrSymbol, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Internal::CursorPage[OpenAI::Models::VectorStores::VectorStoreFile]) end diff --git a/rbi/lib/openai/resources/vector_stores/files.rbi b/rbi/lib/openai/resources/vector_stores/files.rbi index fe472d5a..145d752b 100644 --- a/rbi/lib/openai/resources/vector_stores/files.rbi +++ b/rbi/lib/openai/resources/vector_stores/files.rbi @@ -14,10 +14,10 @@ module OpenAI attributes: T.nilable(T::Hash[Symbol, T.any(String, Float, T::Boolean)]), chunking_strategy: T.any( OpenAI::Models::AutoFileChunkingStrategyParam, - OpenAI::Internal::Util::AnyHash, + OpenAI::Internal::AnyHash, OpenAI::Models::StaticFileChunkingStrategyObjectParam ), - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::VectorStores::VectorStoreFile) end @@ -46,7 +46,7 @@ module OpenAI params( file_id: String, vector_store_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::VectorStores::VectorStoreFile) end @@ -65,7 +65,7 @@ module OpenAI file_id: String, vector_store_id: String, attributes: T.nilable(T::Hash[Symbol, T.any(String, Float, T::Boolean)]), - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::VectorStores::VectorStoreFile) end @@ -93,7 +93,7 @@ module OpenAI filter: OpenAI::Models::VectorStores::FileListParams::Filter::OrSymbol, limit: Integer, order: OpenAI::Models::VectorStores::FileListParams::Order::OrSymbol, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Internal::CursorPage[OpenAI::Models::VectorStores::VectorStoreFile]) end @@ -130,7 +130,7 @@ module OpenAI params( file_id: String, vector_store_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Models::VectorStores::VectorStoreFileDeleted) end @@ -148,7 +148,7 @@ module OpenAI params( file_id: String, vector_store_id: String, - request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::Util::AnyHash)) + request_options: T.nilable(T.any(OpenAI::RequestOptions, OpenAI::Internal::AnyHash)) ) .returns(OpenAI::Internal::Page[OpenAI::Models::VectorStores::FileContentResponse]) end diff --git a/sig/openai/aliases.rbs b/sig/openai/aliases.rbs deleted file mode 100644 index b99ae58a..00000000 --- a/sig/openai/aliases.rbs +++ /dev/null @@ -1,19 +0,0 @@ -module OpenAI - class Unknown = OpenAI::Internal::Type::Unknown - - class BooleanModel = OpenAI::Internal::Type::BooleanModel - - module Enum = OpenAI::Internal::Type::Enum - - module Union = OpenAI::Internal::Type::Union - - class ArrayOf = OpenAI::Internal::Type::ArrayOf - - class HashOf = OpenAI::Internal::Type::HashOf - - class BaseModel = OpenAI::Internal::Type::BaseModel - - type request_parameters = OpenAI::Internal::Type::request_parameters - - module RequestParameters = OpenAI::Internal::Type::RequestParameters -end diff --git a/sig/openai/errors.rbs b/sig/openai/errors.rbs index 8dc6afce..19583ccb 100644 --- a/sig/openai/errors.rbs +++ b/sig/openai/errors.rbs @@ -104,32 +104,4 @@ module OpenAI HTTP_STATUS: Range[Integer] end end - - class Error = OpenAI::Errors::Error - - class ConversionError = OpenAI::Errors::ConversionError - - class APIError = OpenAI::Errors::APIError - - class APIStatusError = OpenAI::Errors::APIStatusError - - class APIConnectionError = OpenAI::Errors::APIConnectionError - - class APITimeoutError = OpenAI::Errors::APITimeoutError - - class BadRequestError = OpenAI::Errors::BadRequestError - - class AuthenticationError = OpenAI::Errors::AuthenticationError - - class PermissionDeniedError = OpenAI::Errors::PermissionDeniedError - - class NotFoundError = OpenAI::Errors::NotFoundError - - class ConflictError = OpenAI::Errors::ConflictError - - class UnprocessableEntityError = OpenAI::Errors::UnprocessableEntityError - - class RateLimitError = OpenAI::Errors::RateLimitError - - class InternalServerError = OpenAI::Errors::InternalServerError end diff --git a/sig/openai/internal.rbs b/sig/openai/internal.rbs new file mode 100644 index 00000000..ff72a6b0 --- /dev/null +++ b/sig/openai/internal.rbs @@ -0,0 +1,4 @@ +module OpenAI + module Internal + end +end diff --git a/sig/openai/internal/type/base_model.rbs b/sig/openai/internal/type/base_model.rbs index fd068625..69aca270 100644 --- a/sig/openai/internal/type/base_model.rbs +++ b/sig/openai/internal/type/base_model.rbs @@ -7,10 +7,10 @@ module OpenAI type known_field = { mode: (:coerce | :dump)?, required: bool, nilable: bool } - def self.known_fields: -> ::Hash[Symbol, (OpenAI::BaseModel::known_field + def self.known_fields: -> ::Hash[Symbol, (OpenAI::Internal::Type::BaseModel::known_field & { type_fn: (^-> OpenAI::Internal::Type::Converter::input) })] - def self.fields: -> ::Hash[Symbol, (OpenAI::BaseModel::known_field + def self.fields: -> ::Hash[Symbol, (OpenAI::Internal::Type::BaseModel::known_field & { type: OpenAI::Internal::Type::Converter::input })] private def self.add_field: ( @@ -52,7 +52,7 @@ module OpenAI def ==: (top other) -> bool def self.coerce: ( - OpenAI::BaseModel | ::Hash[top, top] | top value, + OpenAI::Internal::Type::BaseModel | ::Hash[top, top] | top value, state: OpenAI::Internal::Type::Converter::state ) -> (instance | top) diff --git a/sig/openai/internal/util.rbs b/sig/openai/internal/util.rbs index 21bf07da..07e02294 100644 --- a/sig/openai/internal/util.rbs +++ b/sig/openai/internal/util.rbs @@ -19,8 +19,6 @@ module OpenAI def self?.coerce_hash: (top input) -> (::Hash[top, top] | top) - OMIT: top - def self?.deep_merge_lr: (top lhs, top rhs, ?concat: bool) -> top def self?.deep_merge: ( diff --git a/sig/openai/models/all_models.rbs b/sig/openai/models/all_models.rbs index 77c8f9b8..836c7e7b 100644 --- a/sig/openai/models/all_models.rbs +++ b/sig/openai/models/all_models.rbs @@ -9,7 +9,7 @@ module OpenAI | :"computer-use-preview-2025-03-11" module AllModels - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, OpenAI::Models::chat_model, (:"o1-pro" | :"o1-pro-2025-03-19" diff --git a/sig/openai/models/audio/speech_create_params.rbs b/sig/openai/models/audio/speech_create_params.rbs index 1b50fd38..97aff0df 100644 --- a/sig/openai/models/audio/speech_create_params.rbs +++ b/sig/openai/models/audio/speech_create_params.rbs @@ -12,7 +12,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class SpeechCreateParams < OpenAI::BaseModel + class SpeechCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -51,7 +51,7 @@ module OpenAI type model = String | OpenAI::Models::Audio::speech_model module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, OpenAI::Models::Audio::speech_model] end @@ -71,7 +71,7 @@ module OpenAI | :verse module Voice - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, (:alloy | :ash @@ -101,7 +101,7 @@ module OpenAI type response_format = :mp3 | :opus | :aac | :flac | :wav | :pcm module ResponseFormat - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum MP3: :mp3 OPUS: :opus diff --git a/sig/openai/models/audio/speech_model.rbs b/sig/openai/models/audio/speech_model.rbs index 8fc430d6..1dfa36a8 100644 --- a/sig/openai/models/audio/speech_model.rbs +++ b/sig/openai/models/audio/speech_model.rbs @@ -4,7 +4,7 @@ module OpenAI type speech_model = :"tts-1" | :"tts-1-hd" | :"gpt-4o-mini-tts" module SpeechModel - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TTS_1: :"tts-1" TTS_1_HD: :"tts-1-hd" diff --git a/sig/openai/models/audio/transcription.rbs b/sig/openai/models/audio/transcription.rbs index 3f9bf1d4..12d41fc2 100644 --- a/sig/openai/models/audio/transcription.rbs +++ b/sig/openai/models/audio/transcription.rbs @@ -7,7 +7,7 @@ module OpenAI logprobs: ::Array[OpenAI::Models::Audio::Transcription::Logprob] } - class Transcription < OpenAI::BaseModel + class Transcription < OpenAI::Internal::Type::BaseModel attr_accessor text: String attr_reader logprobs: ::Array[OpenAI::Models::Audio::Transcription::Logprob]? @@ -25,7 +25,7 @@ module OpenAI type logprob = { token: String, bytes: ::Array[Float], logprob: Float } - class Logprob < OpenAI::BaseModel + class Logprob < OpenAI::Internal::Type::BaseModel attr_reader token: String? def token=: (String) -> String diff --git a/sig/openai/models/audio/transcription_create_params.rbs b/sig/openai/models/audio/transcription_create_params.rbs index b46cde10..f0ae2daa 100644 --- a/sig/openai/models/audio/transcription_create_params.rbs +++ b/sig/openai/models/audio/transcription_create_params.rbs @@ -14,7 +14,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class TranscriptionCreateParams < OpenAI::BaseModel + class TranscriptionCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -69,7 +69,7 @@ module OpenAI type model = String | OpenAI::Models::audio_model module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, OpenAI::Models::audio_model] end @@ -77,7 +77,7 @@ module OpenAI type timestamp_granularity = :word | :segment module TimestampGranularity - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum WORD: :word SEGMENT: :segment diff --git a/sig/openai/models/audio/transcription_create_response.rbs b/sig/openai/models/audio/transcription_create_response.rbs index 49eb0f09..f4931f11 100644 --- a/sig/openai/models/audio/transcription_create_response.rbs +++ b/sig/openai/models/audio/transcription_create_response.rbs @@ -6,7 +6,7 @@ module OpenAI | OpenAI::Models::Audio::TranscriptionVerbose module TranscriptionCreateResponse - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Audio::Transcription, OpenAI::Models::Audio::TranscriptionVerbose] end diff --git a/sig/openai/models/audio/transcription_include.rbs b/sig/openai/models/audio/transcription_include.rbs index 349684b6..de5322fa 100644 --- a/sig/openai/models/audio/transcription_include.rbs +++ b/sig/openai/models/audio/transcription_include.rbs @@ -4,7 +4,7 @@ module OpenAI type transcription_include = :logprobs module TranscriptionInclude - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum LOGPROBS: :logprobs diff --git a/sig/openai/models/audio/transcription_segment.rbs b/sig/openai/models/audio/transcription_segment.rbs index 85af1669..b0748151 100644 --- a/sig/openai/models/audio/transcription_segment.rbs +++ b/sig/openai/models/audio/transcription_segment.rbs @@ -15,7 +15,7 @@ module OpenAI tokens: ::Array[Integer] } - class TranscriptionSegment < OpenAI::BaseModel + class TranscriptionSegment < OpenAI::Internal::Type::BaseModel attr_accessor id: Integer attr_accessor avg_logprob: Float diff --git a/sig/openai/models/audio/transcription_stream_event.rbs b/sig/openai/models/audio/transcription_stream_event.rbs index 8441d201..158d4540 100644 --- a/sig/openai/models/audio/transcription_stream_event.rbs +++ b/sig/openai/models/audio/transcription_stream_event.rbs @@ -6,7 +6,7 @@ module OpenAI | OpenAI::Models::Audio::TranscriptionTextDoneEvent module TranscriptionStreamEvent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Audio::TranscriptionTextDeltaEvent, OpenAI::Models::Audio::TranscriptionTextDoneEvent] end diff --git a/sig/openai/models/audio/transcription_text_delta_event.rbs b/sig/openai/models/audio/transcription_text_delta_event.rbs index 373c6aed..c34d5fcd 100644 --- a/sig/openai/models/audio/transcription_text_delta_event.rbs +++ b/sig/openai/models/audio/transcription_text_delta_event.rbs @@ -8,7 +8,7 @@ module OpenAI logprobs: ::Array[OpenAI::Models::Audio::TranscriptionTextDeltaEvent::Logprob] } - class TranscriptionTextDeltaEvent < OpenAI::BaseModel + class TranscriptionTextDeltaEvent < OpenAI::Internal::Type::BaseModel attr_accessor delta: String attr_accessor type: :"transcript.text.delta" @@ -29,7 +29,7 @@ module OpenAI type logprob = { token: String, bytes: ::Array[top], logprob: Float } - class Logprob < OpenAI::BaseModel + class Logprob < OpenAI::Internal::Type::BaseModel attr_reader token: String? def token=: (String) -> String diff --git a/sig/openai/models/audio/transcription_text_done_event.rbs b/sig/openai/models/audio/transcription_text_done_event.rbs index f1f1dd1d..f6216cee 100644 --- a/sig/openai/models/audio/transcription_text_done_event.rbs +++ b/sig/openai/models/audio/transcription_text_done_event.rbs @@ -8,7 +8,7 @@ module OpenAI logprobs: ::Array[OpenAI::Models::Audio::TranscriptionTextDoneEvent::Logprob] } - class TranscriptionTextDoneEvent < OpenAI::BaseModel + class TranscriptionTextDoneEvent < OpenAI::Internal::Type::BaseModel attr_accessor text: String attr_accessor type: :"transcript.text.done" @@ -29,7 +29,7 @@ module OpenAI type logprob = { token: String, bytes: ::Array[top], logprob: Float } - class Logprob < OpenAI::BaseModel + class Logprob < OpenAI::Internal::Type::BaseModel attr_reader token: String? def token=: (String) -> String diff --git a/sig/openai/models/audio/transcription_verbose.rbs b/sig/openai/models/audio/transcription_verbose.rbs index 9546e704..34e8fe7a 100644 --- a/sig/openai/models/audio/transcription_verbose.rbs +++ b/sig/openai/models/audio/transcription_verbose.rbs @@ -10,7 +10,7 @@ module OpenAI words: ::Array[OpenAI::Models::Audio::TranscriptionWord] } - class TranscriptionVerbose < OpenAI::BaseModel + class TranscriptionVerbose < OpenAI::Internal::Type::BaseModel attr_accessor duration: Float attr_accessor language: String diff --git a/sig/openai/models/audio/transcription_word.rbs b/sig/openai/models/audio/transcription_word.rbs index 5b9eee66..e9332bbd 100644 --- a/sig/openai/models/audio/transcription_word.rbs +++ b/sig/openai/models/audio/transcription_word.rbs @@ -3,7 +3,7 @@ module OpenAI module Audio type transcription_word = { end_: Float, start: Float, word: String } - class TranscriptionWord < OpenAI::BaseModel + class TranscriptionWord < OpenAI::Internal::Type::BaseModel attr_accessor end_: Float attr_accessor start: Float diff --git a/sig/openai/models/audio/translation.rbs b/sig/openai/models/audio/translation.rbs index 4a88fe3f..43eeef91 100644 --- a/sig/openai/models/audio/translation.rbs +++ b/sig/openai/models/audio/translation.rbs @@ -3,7 +3,7 @@ module OpenAI module Audio type translation = { text: String } - class Translation < OpenAI::BaseModel + class Translation < OpenAI::Internal::Type::BaseModel attr_accessor text: String def initialize: (text: String) -> void diff --git a/sig/openai/models/audio/translation_create_params.rbs b/sig/openai/models/audio/translation_create_params.rbs index 800cd86d..588ac2cf 100644 --- a/sig/openai/models/audio/translation_create_params.rbs +++ b/sig/openai/models/audio/translation_create_params.rbs @@ -11,7 +11,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class TranslationCreateParams < OpenAI::BaseModel + class TranslationCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -47,7 +47,7 @@ module OpenAI type model = String | OpenAI::Models::audio_model module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, OpenAI::Models::audio_model] end @@ -55,7 +55,7 @@ module OpenAI type response_format = :json | :text | :srt | :verbose_json | :vtt module ResponseFormat - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum JSON: :json TEXT: :text diff --git a/sig/openai/models/audio/translation_create_response.rbs b/sig/openai/models/audio/translation_create_response.rbs index 7b88f273..785dfce2 100644 --- a/sig/openai/models/audio/translation_create_response.rbs +++ b/sig/openai/models/audio/translation_create_response.rbs @@ -6,7 +6,7 @@ module OpenAI | OpenAI::Models::Audio::TranslationVerbose module TranslationCreateResponse - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Audio::Translation, OpenAI::Models::Audio::TranslationVerbose] end diff --git a/sig/openai/models/audio/translation_verbose.rbs b/sig/openai/models/audio/translation_verbose.rbs index 0a805f53..3f69b2ca 100644 --- a/sig/openai/models/audio/translation_verbose.rbs +++ b/sig/openai/models/audio/translation_verbose.rbs @@ -9,7 +9,7 @@ module OpenAI segments: ::Array[OpenAI::Models::Audio::TranscriptionSegment] } - class TranslationVerbose < OpenAI::BaseModel + class TranslationVerbose < OpenAI::Internal::Type::BaseModel attr_accessor duration: Float attr_accessor language: String diff --git a/sig/openai/models/audio_model.rbs b/sig/openai/models/audio_model.rbs index c5a9d284..4a294e19 100644 --- a/sig/openai/models/audio_model.rbs +++ b/sig/openai/models/audio_model.rbs @@ -4,7 +4,7 @@ module OpenAI :"whisper-1" | :"gpt-4o-transcribe" | :"gpt-4o-mini-transcribe" module AudioModel - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum WHISPER_1: :"whisper-1" GPT_4O_TRANSCRIBE: :"gpt-4o-transcribe" diff --git a/sig/openai/models/audio_response_format.rbs b/sig/openai/models/audio_response_format.rbs index 7c91cd99..39091918 100644 --- a/sig/openai/models/audio_response_format.rbs +++ b/sig/openai/models/audio_response_format.rbs @@ -3,7 +3,7 @@ module OpenAI type audio_response_format = :json | :text | :srt | :verbose_json | :vtt module AudioResponseFormat - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum JSON: :json TEXT: :text diff --git a/sig/openai/models/auto_file_chunking_strategy_param.rbs b/sig/openai/models/auto_file_chunking_strategy_param.rbs index 49630c41..a5285e7b 100644 --- a/sig/openai/models/auto_file_chunking_strategy_param.rbs +++ b/sig/openai/models/auto_file_chunking_strategy_param.rbs @@ -2,7 +2,7 @@ module OpenAI module Models type auto_file_chunking_strategy_param = { type: :auto } - class AutoFileChunkingStrategyParam < OpenAI::BaseModel + class AutoFileChunkingStrategyParam < OpenAI::Internal::Type::BaseModel attr_accessor type: :auto def initialize: (?type: :auto) -> void diff --git a/sig/openai/models/batch.rbs b/sig/openai/models/batch.rbs index 46840a64..75a90985 100644 --- a/sig/openai/models/batch.rbs +++ b/sig/openai/models/batch.rbs @@ -24,7 +24,7 @@ module OpenAI request_counts: OpenAI::Models::BatchRequestCounts } - class Batch < OpenAI::BaseModel + class Batch < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor completion_window: String @@ -129,7 +129,7 @@ module OpenAI | :cancelled module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum VALIDATING: :validating FAILED: :failed @@ -146,7 +146,7 @@ module OpenAI type errors = { data: ::Array[OpenAI::Models::BatchError], object: String } - class Errors < OpenAI::BaseModel + class Errors < OpenAI::Internal::Type::BaseModel attr_reader data: ::Array[OpenAI::Models::BatchError]? def data=: ( diff --git a/sig/openai/models/batch_cancel_params.rbs b/sig/openai/models/batch_cancel_params.rbs index fdb4fc66..c8b34f9f 100644 --- a/sig/openai/models/batch_cancel_params.rbs +++ b/sig/openai/models/batch_cancel_params.rbs @@ -2,7 +2,7 @@ module OpenAI module Models type batch_cancel_params = { } & OpenAI::Internal::Type::request_parameters - class BatchCancelParams < OpenAI::BaseModel + class BatchCancelParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/batch_create_params.rbs b/sig/openai/models/batch_create_params.rbs index 19fe80ff..cacafcf2 100644 --- a/sig/openai/models/batch_create_params.rbs +++ b/sig/openai/models/batch_create_params.rbs @@ -9,7 +9,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class BatchCreateParams < OpenAI::BaseModel + class BatchCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -34,7 +34,7 @@ module OpenAI type completion_window = :"24h" module CompletionWindow - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum COMPLETION_WINDOW_24H: :"24h" @@ -48,7 +48,7 @@ module OpenAI | :"/v1/completions" module Endpoint - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum V1_RESPONSES: :"/v1/responses" V1_CHAT_COMPLETIONS: :"/v1/chat/completions" diff --git a/sig/openai/models/batch_error.rbs b/sig/openai/models/batch_error.rbs index b2aadd8e..75e828eb 100644 --- a/sig/openai/models/batch_error.rbs +++ b/sig/openai/models/batch_error.rbs @@ -3,7 +3,7 @@ module OpenAI type batch_error = { code: String, line: Integer?, message: String, param: String? } - class BatchError < OpenAI::BaseModel + class BatchError < OpenAI::Internal::Type::BaseModel attr_reader code: String? def code=: (String) -> String diff --git a/sig/openai/models/batch_list_params.rbs b/sig/openai/models/batch_list_params.rbs index 3ca61260..9cb84ad1 100644 --- a/sig/openai/models/batch_list_params.rbs +++ b/sig/openai/models/batch_list_params.rbs @@ -4,7 +4,7 @@ module OpenAI { after: String, limit: Integer } & OpenAI::Internal::Type::request_parameters - class BatchListParams < OpenAI::BaseModel + class BatchListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/batch_request_counts.rbs b/sig/openai/models/batch_request_counts.rbs index 0020eb38..56860e0e 100644 --- a/sig/openai/models/batch_request_counts.rbs +++ b/sig/openai/models/batch_request_counts.rbs @@ -3,7 +3,7 @@ module OpenAI type batch_request_counts = { completed: Integer, failed: Integer, total: Integer } - class BatchRequestCounts < OpenAI::BaseModel + class BatchRequestCounts < OpenAI::Internal::Type::BaseModel attr_accessor completed: Integer attr_accessor failed: Integer diff --git a/sig/openai/models/batch_retrieve_params.rbs b/sig/openai/models/batch_retrieve_params.rbs index 4b17656d..681750c7 100644 --- a/sig/openai/models/batch_retrieve_params.rbs +++ b/sig/openai/models/batch_retrieve_params.rbs @@ -3,7 +3,7 @@ module OpenAI type batch_retrieve_params = { } & OpenAI::Internal::Type::request_parameters - class BatchRetrieveParams < OpenAI::BaseModel + class BatchRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/beta/assistant.rbs b/sig/openai/models/beta/assistant.rbs index 91f1b5ee..557bcacf 100644 --- a/sig/openai/models/beta/assistant.rbs +++ b/sig/openai/models/beta/assistant.rbs @@ -18,7 +18,7 @@ module OpenAI top_p: Float? } - class Assistant < OpenAI::BaseModel + class Assistant < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor created_at: Integer @@ -69,7 +69,7 @@ module OpenAI file_search: OpenAI::Models::Beta::Assistant::ToolResources::FileSearch } - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel attr_reader code_interpreter: OpenAI::Models::Beta::Assistant::ToolResources::CodeInterpreter? def code_interpreter=: ( @@ -91,7 +91,7 @@ module OpenAI type code_interpreter = { file_ids: ::Array[String] } - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel attr_reader file_ids: ::Array[String]? def file_ids=: (::Array[String]) -> ::Array[String] @@ -103,7 +103,7 @@ module OpenAI type file_search = { vector_store_ids: ::Array[String] } - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel attr_reader vector_store_ids: ::Array[String]? def vector_store_ids=: (::Array[String]) -> ::Array[String] diff --git a/sig/openai/models/beta/assistant_create_params.rbs b/sig/openai/models/beta/assistant_create_params.rbs index 10663a9c..83167a35 100644 --- a/sig/openai/models/beta/assistant_create_params.rbs +++ b/sig/openai/models/beta/assistant_create_params.rbs @@ -17,7 +17,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class AssistantCreateParams < OpenAI::BaseModel + class AssistantCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -67,7 +67,7 @@ module OpenAI type model = String | OpenAI::Models::chat_model module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, OpenAI::Models::chat_model] end @@ -78,7 +78,7 @@ module OpenAI file_search: OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch } - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel attr_reader code_interpreter: OpenAI::Models::Beta::AssistantCreateParams::ToolResources::CodeInterpreter? def code_interpreter=: ( @@ -100,7 +100,7 @@ module OpenAI type code_interpreter = { file_ids: ::Array[String] } - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel attr_reader file_ids: ::Array[String]? def file_ids=: (::Array[String]) -> ::Array[String] @@ -116,7 +116,7 @@ module OpenAI vector_stores: ::Array[OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch::VectorStore] } - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel attr_reader vector_store_ids: ::Array[String]? def vector_store_ids=: (::Array[String]) -> ::Array[String] @@ -141,7 +141,7 @@ module OpenAI metadata: OpenAI::Models::metadata? } - class VectorStore < OpenAI::BaseModel + class VectorStore < OpenAI::Internal::Type::BaseModel attr_reader chunking_strategy: OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch::VectorStore::chunking_strategy? def chunking_strategy=: ( @@ -167,11 +167,11 @@ module OpenAI | OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static module ChunkingStrategy - extend OpenAI::Union + extend OpenAI::Internal::Type::Union type auto = { type: :auto } - class Auto < OpenAI::BaseModel + class Auto < OpenAI::Internal::Type::BaseModel attr_accessor type: :auto def initialize: (?type: :auto) -> void @@ -185,7 +185,7 @@ module OpenAI type: :static } - class Static < OpenAI::BaseModel + class Static < OpenAI::Internal::Type::BaseModel attr_accessor static: OpenAI::Models::Beta::AssistantCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static::Static attr_accessor type: :static @@ -203,7 +203,7 @@ module OpenAI max_chunk_size_tokens: Integer } - class Static < OpenAI::BaseModel + class Static < OpenAI::Internal::Type::BaseModel attr_accessor chunk_overlap_tokens: Integer attr_accessor max_chunk_size_tokens: Integer diff --git a/sig/openai/models/beta/assistant_delete_params.rbs b/sig/openai/models/beta/assistant_delete_params.rbs index 297d2fdb..d6a42d1d 100644 --- a/sig/openai/models/beta/assistant_delete_params.rbs +++ b/sig/openai/models/beta/assistant_delete_params.rbs @@ -4,7 +4,7 @@ module OpenAI type assistant_delete_params = { } & OpenAI::Internal::Type::request_parameters - class AssistantDeleteParams < OpenAI::BaseModel + class AssistantDeleteParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/beta/assistant_deleted.rbs b/sig/openai/models/beta/assistant_deleted.rbs index 39b34057..8c913e60 100644 --- a/sig/openai/models/beta/assistant_deleted.rbs +++ b/sig/openai/models/beta/assistant_deleted.rbs @@ -4,7 +4,7 @@ module OpenAI type assistant_deleted = { id: String, deleted: bool, object: :"assistant.deleted" } - class AssistantDeleted < OpenAI::BaseModel + class AssistantDeleted < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor deleted: bool diff --git a/sig/openai/models/beta/assistant_list_params.rbs b/sig/openai/models/beta/assistant_list_params.rbs index c3040fc8..c0481b7d 100644 --- a/sig/openai/models/beta/assistant_list_params.rbs +++ b/sig/openai/models/beta/assistant_list_params.rbs @@ -10,7 +10,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class AssistantListParams < OpenAI::BaseModel + class AssistantListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -45,7 +45,7 @@ module OpenAI type order = :asc | :desc module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASC: :asc DESC: :desc diff --git a/sig/openai/models/beta/assistant_response_format_option.rbs b/sig/openai/models/beta/assistant_response_format_option.rbs index 3b848b00..e5da5be7 100644 --- a/sig/openai/models/beta/assistant_response_format_option.rbs +++ b/sig/openai/models/beta/assistant_response_format_option.rbs @@ -8,7 +8,7 @@ module OpenAI | OpenAI::Models::ResponseFormatJSONSchema module AssistantResponseFormatOption - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [:auto, OpenAI::Models::ResponseFormatText, OpenAI::Models::ResponseFormatJSONObject, OpenAI::Models::ResponseFormatJSONSchema] end diff --git a/sig/openai/models/beta/assistant_retrieve_params.rbs b/sig/openai/models/beta/assistant_retrieve_params.rbs index d39312a3..53274daa 100644 --- a/sig/openai/models/beta/assistant_retrieve_params.rbs +++ b/sig/openai/models/beta/assistant_retrieve_params.rbs @@ -4,7 +4,7 @@ module OpenAI type assistant_retrieve_params = { } & OpenAI::Internal::Type::request_parameters - class AssistantRetrieveParams < OpenAI::BaseModel + class AssistantRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/beta/assistant_stream_event.rbs b/sig/openai/models/beta/assistant_stream_event.rbs index ae88c40e..915e637d 100644 --- a/sig/openai/models/beta/assistant_stream_event.rbs +++ b/sig/openai/models/beta/assistant_stream_event.rbs @@ -28,7 +28,7 @@ module OpenAI | OpenAI::Models::Beta::AssistantStreamEvent::ErrorEvent module AssistantStreamEvent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union type thread_created = { @@ -37,7 +37,7 @@ module OpenAI enabled: bool } - class ThreadCreated < OpenAI::BaseModel + class ThreadCreated < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Thread attr_accessor event: :"thread.created" @@ -61,7 +61,7 @@ module OpenAI event: :"thread.run.created" } - class ThreadRunCreated < OpenAI::BaseModel + class ThreadRunCreated < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Run attr_accessor event: :"thread.run.created" @@ -80,7 +80,7 @@ module OpenAI event: :"thread.run.queued" } - class ThreadRunQueued < OpenAI::BaseModel + class ThreadRunQueued < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Run attr_accessor event: :"thread.run.queued" @@ -99,7 +99,7 @@ module OpenAI event: :"thread.run.in_progress" } - class ThreadRunInProgress < OpenAI::BaseModel + class ThreadRunInProgress < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Run attr_accessor event: :"thread.run.in_progress" @@ -118,7 +118,7 @@ module OpenAI event: :"thread.run.requires_action" } - class ThreadRunRequiresAction < OpenAI::BaseModel + class ThreadRunRequiresAction < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Run attr_accessor event: :"thread.run.requires_action" @@ -137,7 +137,7 @@ module OpenAI event: :"thread.run.completed" } - class ThreadRunCompleted < OpenAI::BaseModel + class ThreadRunCompleted < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Run attr_accessor event: :"thread.run.completed" @@ -156,7 +156,7 @@ module OpenAI event: :"thread.run.incomplete" } - class ThreadRunIncomplete < OpenAI::BaseModel + class ThreadRunIncomplete < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Run attr_accessor event: :"thread.run.incomplete" @@ -175,7 +175,7 @@ module OpenAI event: :"thread.run.failed" } - class ThreadRunFailed < OpenAI::BaseModel + class ThreadRunFailed < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Run attr_accessor event: :"thread.run.failed" @@ -194,7 +194,7 @@ module OpenAI event: :"thread.run.cancelling" } - class ThreadRunCancelling < OpenAI::BaseModel + class ThreadRunCancelling < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Run attr_accessor event: :"thread.run.cancelling" @@ -213,7 +213,7 @@ module OpenAI event: :"thread.run.cancelled" } - class ThreadRunCancelled < OpenAI::BaseModel + class ThreadRunCancelled < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Run attr_accessor event: :"thread.run.cancelled" @@ -232,7 +232,7 @@ module OpenAI event: :"thread.run.expired" } - class ThreadRunExpired < OpenAI::BaseModel + class ThreadRunExpired < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Run attr_accessor event: :"thread.run.expired" @@ -251,7 +251,7 @@ module OpenAI event: :"thread.run.step.created" } - class ThreadRunStepCreated < OpenAI::BaseModel + class ThreadRunStepCreated < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Runs::RunStep attr_accessor event: :"thread.run.step.created" @@ -270,7 +270,7 @@ module OpenAI event: :"thread.run.step.in_progress" } - class ThreadRunStepInProgress < OpenAI::BaseModel + class ThreadRunStepInProgress < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Runs::RunStep attr_accessor event: :"thread.run.step.in_progress" @@ -289,7 +289,7 @@ module OpenAI event: :"thread.run.step.delta" } - class ThreadRunStepDelta < OpenAI::BaseModel + class ThreadRunStepDelta < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Runs::RunStepDeltaEvent attr_accessor event: :"thread.run.step.delta" @@ -308,7 +308,7 @@ module OpenAI event: :"thread.run.step.completed" } - class ThreadRunStepCompleted < OpenAI::BaseModel + class ThreadRunStepCompleted < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Runs::RunStep attr_accessor event: :"thread.run.step.completed" @@ -327,7 +327,7 @@ module OpenAI event: :"thread.run.step.failed" } - class ThreadRunStepFailed < OpenAI::BaseModel + class ThreadRunStepFailed < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Runs::RunStep attr_accessor event: :"thread.run.step.failed" @@ -346,7 +346,7 @@ module OpenAI event: :"thread.run.step.cancelled" } - class ThreadRunStepCancelled < OpenAI::BaseModel + class ThreadRunStepCancelled < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Runs::RunStep attr_accessor event: :"thread.run.step.cancelled" @@ -365,7 +365,7 @@ module OpenAI event: :"thread.run.step.expired" } - class ThreadRunStepExpired < OpenAI::BaseModel + class ThreadRunStepExpired < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Runs::RunStep attr_accessor event: :"thread.run.step.expired" @@ -384,7 +384,7 @@ module OpenAI event: :"thread.message.created" } - class ThreadMessageCreated < OpenAI::BaseModel + class ThreadMessageCreated < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Message attr_accessor event: :"thread.message.created" @@ -403,7 +403,7 @@ module OpenAI event: :"thread.message.in_progress" } - class ThreadMessageInProgress < OpenAI::BaseModel + class ThreadMessageInProgress < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Message attr_accessor event: :"thread.message.in_progress" @@ -422,7 +422,7 @@ module OpenAI event: :"thread.message.delta" } - class ThreadMessageDelta < OpenAI::BaseModel + class ThreadMessageDelta < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::MessageDeltaEvent attr_accessor event: :"thread.message.delta" @@ -441,7 +441,7 @@ module OpenAI event: :"thread.message.completed" } - class ThreadMessageCompleted < OpenAI::BaseModel + class ThreadMessageCompleted < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Message attr_accessor event: :"thread.message.completed" @@ -460,7 +460,7 @@ module OpenAI event: :"thread.message.incomplete" } - class ThreadMessageIncomplete < OpenAI::BaseModel + class ThreadMessageIncomplete < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Message attr_accessor event: :"thread.message.incomplete" @@ -475,7 +475,7 @@ module OpenAI type error_event = { data: OpenAI::Models::ErrorObject, event: :error } - class ErrorEvent < OpenAI::BaseModel + class ErrorEvent < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::ErrorObject attr_accessor event: :error diff --git a/sig/openai/models/beta/assistant_tool.rbs b/sig/openai/models/beta/assistant_tool.rbs index d3a43fed..abdfdcff 100644 --- a/sig/openai/models/beta/assistant_tool.rbs +++ b/sig/openai/models/beta/assistant_tool.rbs @@ -7,7 +7,7 @@ module OpenAI | OpenAI::Models::Beta::FunctionTool module AssistantTool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Beta::CodeInterpreterTool, OpenAI::Models::Beta::FileSearchTool, OpenAI::Models::Beta::FunctionTool] end diff --git a/sig/openai/models/beta/assistant_tool_choice.rbs b/sig/openai/models/beta/assistant_tool_choice.rbs index 69e6e1b5..92b87de5 100644 --- a/sig/openai/models/beta/assistant_tool_choice.rbs +++ b/sig/openai/models/beta/assistant_tool_choice.rbs @@ -7,7 +7,7 @@ module OpenAI function: OpenAI::Models::Beta::AssistantToolChoiceFunction } - class AssistantToolChoice < OpenAI::BaseModel + class AssistantToolChoice < OpenAI::Internal::Type::BaseModel attr_accessor type: OpenAI::Models::Beta::AssistantToolChoice::type_ attr_reader function: OpenAI::Models::Beta::AssistantToolChoiceFunction? @@ -26,7 +26,7 @@ module OpenAI type type_ = :function | :code_interpreter | :file_search module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum FUNCTION: :function CODE_INTERPRETER: :code_interpreter diff --git a/sig/openai/models/beta/assistant_tool_choice_function.rbs b/sig/openai/models/beta/assistant_tool_choice_function.rbs index ab112b3d..bd41ba31 100644 --- a/sig/openai/models/beta/assistant_tool_choice_function.rbs +++ b/sig/openai/models/beta/assistant_tool_choice_function.rbs @@ -3,7 +3,7 @@ module OpenAI module Beta type assistant_tool_choice_function = { name: String } - class AssistantToolChoiceFunction < OpenAI::BaseModel + class AssistantToolChoiceFunction < OpenAI::Internal::Type::BaseModel attr_accessor name: String def initialize: (name: String) -> void diff --git a/sig/openai/models/beta/assistant_tool_choice_option.rbs b/sig/openai/models/beta/assistant_tool_choice_option.rbs index 4b1d9ff3..54a6f6c5 100644 --- a/sig/openai/models/beta/assistant_tool_choice_option.rbs +++ b/sig/openai/models/beta/assistant_tool_choice_option.rbs @@ -6,12 +6,12 @@ module OpenAI | OpenAI::Models::Beta::AssistantToolChoice module AssistantToolChoiceOption - extend OpenAI::Union + extend OpenAI::Internal::Type::Union type auto = :none | :auto | :required module Auto - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum NONE: :none AUTO: :auto diff --git a/sig/openai/models/beta/assistant_update_params.rbs b/sig/openai/models/beta/assistant_update_params.rbs index f84dde39..de95efb8 100644 --- a/sig/openai/models/beta/assistant_update_params.rbs +++ b/sig/openai/models/beta/assistant_update_params.rbs @@ -17,7 +17,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class AssistantUpdateParams < OpenAI::BaseModel + class AssistantUpdateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -102,7 +102,7 @@ module OpenAI | :"gpt-3.5-turbo-16k-0613" module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, (:"o3-mini" | :"o3-mini-2025-01-31" @@ -173,7 +173,7 @@ module OpenAI file_search: OpenAI::Models::Beta::AssistantUpdateParams::ToolResources::FileSearch } - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel attr_reader code_interpreter: OpenAI::Models::Beta::AssistantUpdateParams::ToolResources::CodeInterpreter? def code_interpreter=: ( @@ -195,7 +195,7 @@ module OpenAI type code_interpreter = { file_ids: ::Array[String] } - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel attr_reader file_ids: ::Array[String]? def file_ids=: (::Array[String]) -> ::Array[String] @@ -207,7 +207,7 @@ module OpenAI type file_search = { vector_store_ids: ::Array[String] } - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel attr_reader vector_store_ids: ::Array[String]? def vector_store_ids=: (::Array[String]) -> ::Array[String] diff --git a/sig/openai/models/beta/code_interpreter_tool.rbs b/sig/openai/models/beta/code_interpreter_tool.rbs index 3855b043..f986bb74 100644 --- a/sig/openai/models/beta/code_interpreter_tool.rbs +++ b/sig/openai/models/beta/code_interpreter_tool.rbs @@ -3,7 +3,7 @@ module OpenAI module Beta type code_interpreter_tool = { type: :code_interpreter } - class CodeInterpreterTool < OpenAI::BaseModel + class CodeInterpreterTool < OpenAI::Internal::Type::BaseModel attr_accessor type: :code_interpreter def initialize: (?type: :code_interpreter) -> void diff --git a/sig/openai/models/beta/file_search_tool.rbs b/sig/openai/models/beta/file_search_tool.rbs index c18305c9..f3bec757 100644 --- a/sig/openai/models/beta/file_search_tool.rbs +++ b/sig/openai/models/beta/file_search_tool.rbs @@ -7,7 +7,7 @@ module OpenAI file_search: OpenAI::Models::Beta::FileSearchTool::FileSearch } - class FileSearchTool < OpenAI::BaseModel + class FileSearchTool < OpenAI::Internal::Type::BaseModel attr_accessor type: :file_search attr_reader file_search: OpenAI::Models::Beta::FileSearchTool::FileSearch? @@ -29,7 +29,7 @@ module OpenAI ranking_options: OpenAI::Models::Beta::FileSearchTool::FileSearch::RankingOptions } - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel attr_reader max_num_results: Integer? def max_num_results=: (Integer) -> Integer @@ -53,7 +53,7 @@ module OpenAI ranker: OpenAI::Models::Beta::FileSearchTool::FileSearch::RankingOptions::ranker } - class RankingOptions < OpenAI::BaseModel + class RankingOptions < OpenAI::Internal::Type::BaseModel attr_accessor score_threshold: Float attr_reader ranker: OpenAI::Models::Beta::FileSearchTool::FileSearch::RankingOptions::ranker? @@ -72,7 +72,7 @@ module OpenAI type ranker = :auto | :default_2024_08_21 module Ranker - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO: :auto DEFAULT_2024_08_21: :default_2024_08_21 diff --git a/sig/openai/models/beta/function_tool.rbs b/sig/openai/models/beta/function_tool.rbs index 72724066..cdb1900f 100644 --- a/sig/openai/models/beta/function_tool.rbs +++ b/sig/openai/models/beta/function_tool.rbs @@ -4,7 +4,7 @@ module OpenAI type function_tool = { function: OpenAI::Models::FunctionDefinition, type: :function } - class FunctionTool < OpenAI::BaseModel + class FunctionTool < OpenAI::Internal::Type::BaseModel attr_accessor function: OpenAI::Models::FunctionDefinition attr_accessor type: :function diff --git a/sig/openai/models/beta/message_stream_event.rbs b/sig/openai/models/beta/message_stream_event.rbs index fb0a513a..80b9ef55 100644 --- a/sig/openai/models/beta/message_stream_event.rbs +++ b/sig/openai/models/beta/message_stream_event.rbs @@ -9,7 +9,7 @@ module OpenAI | OpenAI::Models::Beta::MessageStreamEvent::ThreadMessageIncomplete module MessageStreamEvent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union type thread_message_created = { @@ -17,7 +17,7 @@ module OpenAI event: :"thread.message.created" } - class ThreadMessageCreated < OpenAI::BaseModel + class ThreadMessageCreated < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Message attr_accessor event: :"thread.message.created" @@ -36,7 +36,7 @@ module OpenAI event: :"thread.message.in_progress" } - class ThreadMessageInProgress < OpenAI::BaseModel + class ThreadMessageInProgress < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Message attr_accessor event: :"thread.message.in_progress" @@ -55,7 +55,7 @@ module OpenAI event: :"thread.message.delta" } - class ThreadMessageDelta < OpenAI::BaseModel + class ThreadMessageDelta < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::MessageDeltaEvent attr_accessor event: :"thread.message.delta" @@ -74,7 +74,7 @@ module OpenAI event: :"thread.message.completed" } - class ThreadMessageCompleted < OpenAI::BaseModel + class ThreadMessageCompleted < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Message attr_accessor event: :"thread.message.completed" @@ -93,7 +93,7 @@ module OpenAI event: :"thread.message.incomplete" } - class ThreadMessageIncomplete < OpenAI::BaseModel + class ThreadMessageIncomplete < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Message attr_accessor event: :"thread.message.incomplete" diff --git a/sig/openai/models/beta/run_step_stream_event.rbs b/sig/openai/models/beta/run_step_stream_event.rbs index 027b49a0..6cd64a35 100644 --- a/sig/openai/models/beta/run_step_stream_event.rbs +++ b/sig/openai/models/beta/run_step_stream_event.rbs @@ -11,7 +11,7 @@ module OpenAI | OpenAI::Models::Beta::RunStepStreamEvent::ThreadRunStepExpired module RunStepStreamEvent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union type thread_run_step_created = { @@ -19,7 +19,7 @@ module OpenAI event: :"thread.run.step.created" } - class ThreadRunStepCreated < OpenAI::BaseModel + class ThreadRunStepCreated < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Runs::RunStep attr_accessor event: :"thread.run.step.created" @@ -38,7 +38,7 @@ module OpenAI event: :"thread.run.step.in_progress" } - class ThreadRunStepInProgress < OpenAI::BaseModel + class ThreadRunStepInProgress < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Runs::RunStep attr_accessor event: :"thread.run.step.in_progress" @@ -57,7 +57,7 @@ module OpenAI event: :"thread.run.step.delta" } - class ThreadRunStepDelta < OpenAI::BaseModel + class ThreadRunStepDelta < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Runs::RunStepDeltaEvent attr_accessor event: :"thread.run.step.delta" @@ -76,7 +76,7 @@ module OpenAI event: :"thread.run.step.completed" } - class ThreadRunStepCompleted < OpenAI::BaseModel + class ThreadRunStepCompleted < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Runs::RunStep attr_accessor event: :"thread.run.step.completed" @@ -95,7 +95,7 @@ module OpenAI event: :"thread.run.step.failed" } - class ThreadRunStepFailed < OpenAI::BaseModel + class ThreadRunStepFailed < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Runs::RunStep attr_accessor event: :"thread.run.step.failed" @@ -114,7 +114,7 @@ module OpenAI event: :"thread.run.step.cancelled" } - class ThreadRunStepCancelled < OpenAI::BaseModel + class ThreadRunStepCancelled < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Runs::RunStep attr_accessor event: :"thread.run.step.cancelled" @@ -133,7 +133,7 @@ module OpenAI event: :"thread.run.step.expired" } - class ThreadRunStepExpired < OpenAI::BaseModel + class ThreadRunStepExpired < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Runs::RunStep attr_accessor event: :"thread.run.step.expired" diff --git a/sig/openai/models/beta/run_stream_event.rbs b/sig/openai/models/beta/run_stream_event.rbs index 87e168ab..3e2604b4 100644 --- a/sig/openai/models/beta/run_stream_event.rbs +++ b/sig/openai/models/beta/run_stream_event.rbs @@ -14,7 +14,7 @@ module OpenAI | OpenAI::Models::Beta::RunStreamEvent::ThreadRunExpired module RunStreamEvent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union type thread_run_created = { @@ -22,7 +22,7 @@ module OpenAI event: :"thread.run.created" } - class ThreadRunCreated < OpenAI::BaseModel + class ThreadRunCreated < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Run attr_accessor event: :"thread.run.created" @@ -41,7 +41,7 @@ module OpenAI event: :"thread.run.queued" } - class ThreadRunQueued < OpenAI::BaseModel + class ThreadRunQueued < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Run attr_accessor event: :"thread.run.queued" @@ -60,7 +60,7 @@ module OpenAI event: :"thread.run.in_progress" } - class ThreadRunInProgress < OpenAI::BaseModel + class ThreadRunInProgress < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Run attr_accessor event: :"thread.run.in_progress" @@ -79,7 +79,7 @@ module OpenAI event: :"thread.run.requires_action" } - class ThreadRunRequiresAction < OpenAI::BaseModel + class ThreadRunRequiresAction < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Run attr_accessor event: :"thread.run.requires_action" @@ -98,7 +98,7 @@ module OpenAI event: :"thread.run.completed" } - class ThreadRunCompleted < OpenAI::BaseModel + class ThreadRunCompleted < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Run attr_accessor event: :"thread.run.completed" @@ -117,7 +117,7 @@ module OpenAI event: :"thread.run.incomplete" } - class ThreadRunIncomplete < OpenAI::BaseModel + class ThreadRunIncomplete < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Run attr_accessor event: :"thread.run.incomplete" @@ -136,7 +136,7 @@ module OpenAI event: :"thread.run.failed" } - class ThreadRunFailed < OpenAI::BaseModel + class ThreadRunFailed < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Run attr_accessor event: :"thread.run.failed" @@ -155,7 +155,7 @@ module OpenAI event: :"thread.run.cancelling" } - class ThreadRunCancelling < OpenAI::BaseModel + class ThreadRunCancelling < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Run attr_accessor event: :"thread.run.cancelling" @@ -174,7 +174,7 @@ module OpenAI event: :"thread.run.cancelled" } - class ThreadRunCancelled < OpenAI::BaseModel + class ThreadRunCancelled < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Run attr_accessor event: :"thread.run.cancelled" @@ -193,7 +193,7 @@ module OpenAI event: :"thread.run.expired" } - class ThreadRunExpired < OpenAI::BaseModel + class ThreadRunExpired < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Threads::Run attr_accessor event: :"thread.run.expired" diff --git a/sig/openai/models/beta/thread.rbs b/sig/openai/models/beta/thread.rbs index 1fdf76c8..e0418d02 100644 --- a/sig/openai/models/beta/thread.rbs +++ b/sig/openai/models/beta/thread.rbs @@ -10,7 +10,7 @@ module OpenAI tool_resources: OpenAI::Models::Beta::Thread::ToolResources? } - class Thread < OpenAI::BaseModel + class Thread < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor created_at: Integer @@ -37,7 +37,7 @@ module OpenAI file_search: OpenAI::Models::Beta::Thread::ToolResources::FileSearch } - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel attr_reader code_interpreter: OpenAI::Models::Beta::Thread::ToolResources::CodeInterpreter? def code_interpreter=: ( @@ -59,7 +59,7 @@ module OpenAI type code_interpreter = { file_ids: ::Array[String] } - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel attr_reader file_ids: ::Array[String]? def file_ids=: (::Array[String]) -> ::Array[String] @@ -71,7 +71,7 @@ module OpenAI type file_search = { vector_store_ids: ::Array[String] } - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel attr_reader vector_store_ids: ::Array[String]? def vector_store_ids=: (::Array[String]) -> ::Array[String] diff --git a/sig/openai/models/beta/thread_create_and_run_params.rbs b/sig/openai/models/beta/thread_create_and_run_params.rbs index 4ca7db99..aeb818ab 100644 --- a/sig/openai/models/beta/thread_create_and_run_params.rbs +++ b/sig/openai/models/beta/thread_create_and_run_params.rbs @@ -21,7 +21,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class ThreadCreateAndRunParams < OpenAI::BaseModel + class ThreadCreateAndRunParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -85,7 +85,7 @@ module OpenAI type model = String | OpenAI::Models::chat_model module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, OpenAI::Models::chat_model] end @@ -97,7 +97,7 @@ module OpenAI tool_resources: OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources? } - class Thread < OpenAI::BaseModel + class Thread < OpenAI::Internal::Type::BaseModel attr_reader messages: ::Array[OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::Message]? def messages=: ( @@ -124,7 +124,7 @@ module OpenAI metadata: OpenAI::Models::metadata? } - class Message < OpenAI::BaseModel + class Message < OpenAI::Internal::Type::BaseModel attr_accessor content: OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::Message::content attr_accessor role: OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::Message::role @@ -147,7 +147,7 @@ module OpenAI | ::Array[OpenAI::Models::Beta::Threads::message_content_part_param] module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, ::Array[OpenAI::Models::Beta::Threads::message_content_part_param]] @@ -160,7 +160,7 @@ module OpenAI type role = :user | :assistant module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum USER: :user ASSISTANT: :assistant @@ -174,7 +174,7 @@ module OpenAI tools: ::Array[OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::Message::Attachment::tool] } - class Attachment < OpenAI::BaseModel + class Attachment < OpenAI::Internal::Type::BaseModel attr_reader file_id: String? def file_id=: (String) -> String @@ -197,11 +197,11 @@ module OpenAI | OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::Message::Attachment::Tool::FileSearch module Tool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union type file_search = { type: :file_search } - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel attr_accessor type: :file_search def initialize: (?type: :file_search) -> void @@ -220,7 +220,7 @@ module OpenAI file_search: OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch } - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel attr_reader code_interpreter: OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::CodeInterpreter? def code_interpreter=: ( @@ -242,7 +242,7 @@ module OpenAI type code_interpreter = { file_ids: ::Array[String] } - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel attr_reader file_ids: ::Array[String]? def file_ids=: (::Array[String]) -> ::Array[String] @@ -258,7 +258,7 @@ module OpenAI vector_stores: ::Array[OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch::VectorStore] } - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel attr_reader vector_store_ids: ::Array[String]? def vector_store_ids=: (::Array[String]) -> ::Array[String] @@ -283,7 +283,7 @@ module OpenAI metadata: OpenAI::Models::metadata? } - class VectorStore < OpenAI::BaseModel + class VectorStore < OpenAI::Internal::Type::BaseModel attr_reader chunking_strategy: OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch::VectorStore::chunking_strategy? def chunking_strategy=: ( @@ -309,11 +309,11 @@ module OpenAI | OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static module ChunkingStrategy - extend OpenAI::Union + extend OpenAI::Internal::Type::Union type auto = { type: :auto } - class Auto < OpenAI::BaseModel + class Auto < OpenAI::Internal::Type::BaseModel attr_accessor type: :auto def initialize: (?type: :auto) -> void @@ -327,7 +327,7 @@ module OpenAI type: :static } - class Static < OpenAI::BaseModel + class Static < OpenAI::Internal::Type::BaseModel attr_accessor static: OpenAI::Models::Beta::ThreadCreateAndRunParams::Thread::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static::Static attr_accessor type: :static @@ -345,7 +345,7 @@ module OpenAI max_chunk_size_tokens: Integer } - class Static < OpenAI::BaseModel + class Static < OpenAI::Internal::Type::BaseModel attr_accessor chunk_overlap_tokens: Integer attr_accessor max_chunk_size_tokens: Integer @@ -372,7 +372,7 @@ module OpenAI file_search: OpenAI::Models::Beta::ThreadCreateAndRunParams::ToolResources::FileSearch } - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel attr_reader code_interpreter: OpenAI::Models::Beta::ThreadCreateAndRunParams::ToolResources::CodeInterpreter? def code_interpreter=: ( @@ -394,7 +394,7 @@ module OpenAI type code_interpreter = { file_ids: ::Array[String] } - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel attr_reader file_ids: ::Array[String]? def file_ids=: (::Array[String]) -> ::Array[String] @@ -406,7 +406,7 @@ module OpenAI type file_search = { vector_store_ids: ::Array[String] } - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel attr_reader vector_store_ids: ::Array[String]? def vector_store_ids=: (::Array[String]) -> ::Array[String] @@ -423,7 +423,7 @@ module OpenAI | OpenAI::Models::Beta::FunctionTool module Tool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Beta::CodeInterpreterTool, OpenAI::Models::Beta::FileSearchTool, OpenAI::Models::Beta::FunctionTool] end @@ -434,7 +434,7 @@ module OpenAI last_messages: Integer? } - class TruncationStrategy < OpenAI::BaseModel + class TruncationStrategy < OpenAI::Internal::Type::BaseModel attr_accessor type: OpenAI::Models::Beta::ThreadCreateAndRunParams::TruncationStrategy::type_ attr_accessor last_messages: Integer? @@ -449,7 +449,7 @@ module OpenAI type type_ = :auto | :last_messages module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO: :auto LAST_MESSAGES: :last_messages diff --git a/sig/openai/models/beta/thread_create_params.rbs b/sig/openai/models/beta/thread_create_params.rbs index 6caf45d4..96d0d366 100644 --- a/sig/openai/models/beta/thread_create_params.rbs +++ b/sig/openai/models/beta/thread_create_params.rbs @@ -9,7 +9,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class ThreadCreateParams < OpenAI::BaseModel + class ThreadCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -40,7 +40,7 @@ module OpenAI metadata: OpenAI::Models::metadata? } - class Message < OpenAI::BaseModel + class Message < OpenAI::Internal::Type::BaseModel attr_accessor content: OpenAI::Models::Beta::ThreadCreateParams::Message::content attr_accessor role: OpenAI::Models::Beta::ThreadCreateParams::Message::role @@ -63,7 +63,7 @@ module OpenAI | ::Array[OpenAI::Models::Beta::Threads::message_content_part_param] module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, ::Array[OpenAI::Models::Beta::Threads::message_content_part_param]] @@ -76,7 +76,7 @@ module OpenAI type role = :user | :assistant module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum USER: :user ASSISTANT: :assistant @@ -90,7 +90,7 @@ module OpenAI tools: ::Array[OpenAI::Models::Beta::ThreadCreateParams::Message::Attachment::tool] } - class Attachment < OpenAI::BaseModel + class Attachment < OpenAI::Internal::Type::BaseModel attr_reader file_id: String? def file_id=: (String) -> String @@ -113,11 +113,11 @@ module OpenAI | OpenAI::Models::Beta::ThreadCreateParams::Message::Attachment::Tool::FileSearch module Tool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union type file_search = { type: :file_search } - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel attr_accessor type: :file_search def initialize: (?type: :file_search) -> void @@ -136,7 +136,7 @@ module OpenAI file_search: OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch } - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel attr_reader code_interpreter: OpenAI::Models::Beta::ThreadCreateParams::ToolResources::CodeInterpreter? def code_interpreter=: ( @@ -158,7 +158,7 @@ module OpenAI type code_interpreter = { file_ids: ::Array[String] } - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel attr_reader file_ids: ::Array[String]? def file_ids=: (::Array[String]) -> ::Array[String] @@ -174,7 +174,7 @@ module OpenAI vector_stores: ::Array[OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch::VectorStore] } - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel attr_reader vector_store_ids: ::Array[String]? def vector_store_ids=: (::Array[String]) -> ::Array[String] @@ -199,7 +199,7 @@ module OpenAI metadata: OpenAI::Models::metadata? } - class VectorStore < OpenAI::BaseModel + class VectorStore < OpenAI::Internal::Type::BaseModel attr_reader chunking_strategy: OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch::VectorStore::chunking_strategy? def chunking_strategy=: ( @@ -225,11 +225,11 @@ module OpenAI | OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static module ChunkingStrategy - extend OpenAI::Union + extend OpenAI::Internal::Type::Union type auto = { type: :auto } - class Auto < OpenAI::BaseModel + class Auto < OpenAI::Internal::Type::BaseModel attr_accessor type: :auto def initialize: (?type: :auto) -> void @@ -243,7 +243,7 @@ module OpenAI type: :static } - class Static < OpenAI::BaseModel + class Static < OpenAI::Internal::Type::BaseModel attr_accessor static: OpenAI::Models::Beta::ThreadCreateParams::ToolResources::FileSearch::VectorStore::ChunkingStrategy::Static::Static attr_accessor type: :static @@ -261,7 +261,7 @@ module OpenAI max_chunk_size_tokens: Integer } - class Static < OpenAI::BaseModel + class Static < OpenAI::Internal::Type::BaseModel attr_accessor chunk_overlap_tokens: Integer attr_accessor max_chunk_size_tokens: Integer diff --git a/sig/openai/models/beta/thread_delete_params.rbs b/sig/openai/models/beta/thread_delete_params.rbs index 27afc166..485a121a 100644 --- a/sig/openai/models/beta/thread_delete_params.rbs +++ b/sig/openai/models/beta/thread_delete_params.rbs @@ -4,7 +4,7 @@ module OpenAI type thread_delete_params = { } & OpenAI::Internal::Type::request_parameters - class ThreadDeleteParams < OpenAI::BaseModel + class ThreadDeleteParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/beta/thread_deleted.rbs b/sig/openai/models/beta/thread_deleted.rbs index c3b1ff2b..1734e860 100644 --- a/sig/openai/models/beta/thread_deleted.rbs +++ b/sig/openai/models/beta/thread_deleted.rbs @@ -4,7 +4,7 @@ module OpenAI type thread_deleted = { id: String, deleted: bool, object: :"thread.deleted" } - class ThreadDeleted < OpenAI::BaseModel + class ThreadDeleted < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor deleted: bool diff --git a/sig/openai/models/beta/thread_retrieve_params.rbs b/sig/openai/models/beta/thread_retrieve_params.rbs index 4db38c39..90d1a84a 100644 --- a/sig/openai/models/beta/thread_retrieve_params.rbs +++ b/sig/openai/models/beta/thread_retrieve_params.rbs @@ -4,7 +4,7 @@ module OpenAI type thread_retrieve_params = { } & OpenAI::Internal::Type::request_parameters - class ThreadRetrieveParams < OpenAI::BaseModel + class ThreadRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/beta/thread_stream_event.rbs b/sig/openai/models/beta/thread_stream_event.rbs index 672ef1d6..74c61286 100644 --- a/sig/openai/models/beta/thread_stream_event.rbs +++ b/sig/openai/models/beta/thread_stream_event.rbs @@ -8,7 +8,7 @@ module OpenAI enabled: bool } - class ThreadStreamEvent < OpenAI::BaseModel + class ThreadStreamEvent < OpenAI::Internal::Type::BaseModel attr_accessor data: OpenAI::Models::Beta::Thread attr_accessor event: :"thread.created" diff --git a/sig/openai/models/beta/thread_update_params.rbs b/sig/openai/models/beta/thread_update_params.rbs index 83b3ecc8..81883e74 100644 --- a/sig/openai/models/beta/thread_update_params.rbs +++ b/sig/openai/models/beta/thread_update_params.rbs @@ -8,7 +8,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class ThreadUpdateParams < OpenAI::BaseModel + class ThreadUpdateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -30,7 +30,7 @@ module OpenAI file_search: OpenAI::Models::Beta::ThreadUpdateParams::ToolResources::FileSearch } - class ToolResources < OpenAI::BaseModel + class ToolResources < OpenAI::Internal::Type::BaseModel attr_reader code_interpreter: OpenAI::Models::Beta::ThreadUpdateParams::ToolResources::CodeInterpreter? def code_interpreter=: ( @@ -52,7 +52,7 @@ module OpenAI type code_interpreter = { file_ids: ::Array[String] } - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel attr_reader file_ids: ::Array[String]? def file_ids=: (::Array[String]) -> ::Array[String] @@ -64,7 +64,7 @@ module OpenAI type file_search = { vector_store_ids: ::Array[String] } - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel attr_reader vector_store_ids: ::Array[String]? def vector_store_ids=: (::Array[String]) -> ::Array[String] diff --git a/sig/openai/models/beta/threads/annotation.rbs b/sig/openai/models/beta/threads/annotation.rbs index 00961486..72c5b0ec 100644 --- a/sig/openai/models/beta/threads/annotation.rbs +++ b/sig/openai/models/beta/threads/annotation.rbs @@ -7,7 +7,7 @@ module OpenAI | OpenAI::Models::Beta::Threads::FilePathAnnotation module Annotation - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Beta::Threads::FileCitationAnnotation, OpenAI::Models::Beta::Threads::FilePathAnnotation] end diff --git a/sig/openai/models/beta/threads/annotation_delta.rbs b/sig/openai/models/beta/threads/annotation_delta.rbs index 84b0d6a2..0840b591 100644 --- a/sig/openai/models/beta/threads/annotation_delta.rbs +++ b/sig/openai/models/beta/threads/annotation_delta.rbs @@ -7,7 +7,7 @@ module OpenAI | OpenAI::Models::Beta::Threads::FilePathDeltaAnnotation module AnnotationDelta - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Beta::Threads::FileCitationDeltaAnnotation, OpenAI::Models::Beta::Threads::FilePathDeltaAnnotation] end diff --git a/sig/openai/models/beta/threads/file_citation_annotation.rbs b/sig/openai/models/beta/threads/file_citation_annotation.rbs index da859605..b22f592c 100644 --- a/sig/openai/models/beta/threads/file_citation_annotation.rbs +++ b/sig/openai/models/beta/threads/file_citation_annotation.rbs @@ -11,7 +11,7 @@ module OpenAI type: :file_citation } - class FileCitationAnnotation < OpenAI::BaseModel + class FileCitationAnnotation < OpenAI::Internal::Type::BaseModel attr_accessor end_index: Integer attr_accessor file_citation: OpenAI::Models::Beta::Threads::FileCitationAnnotation::FileCitation @@ -34,7 +34,7 @@ module OpenAI type file_citation = { file_id: String } - class FileCitation < OpenAI::BaseModel + class FileCitation < OpenAI::Internal::Type::BaseModel attr_accessor file_id: String def initialize: (file_id: String) -> void diff --git a/sig/openai/models/beta/threads/file_citation_delta_annotation.rbs b/sig/openai/models/beta/threads/file_citation_delta_annotation.rbs index 09994a1b..eff94443 100644 --- a/sig/openai/models/beta/threads/file_citation_delta_annotation.rbs +++ b/sig/openai/models/beta/threads/file_citation_delta_annotation.rbs @@ -12,7 +12,7 @@ module OpenAI text: String } - class FileCitationDeltaAnnotation < OpenAI::BaseModel + class FileCitationDeltaAnnotation < OpenAI::Internal::Type::BaseModel attr_accessor index: Integer attr_accessor type: :file_citation @@ -48,7 +48,7 @@ module OpenAI type file_citation = { file_id: String, quote: String } - class FileCitation < OpenAI::BaseModel + class FileCitation < OpenAI::Internal::Type::BaseModel attr_reader file_id: String? def file_id=: (String) -> String diff --git a/sig/openai/models/beta/threads/file_path_annotation.rbs b/sig/openai/models/beta/threads/file_path_annotation.rbs index d3e7acc5..4234a0fd 100644 --- a/sig/openai/models/beta/threads/file_path_annotation.rbs +++ b/sig/openai/models/beta/threads/file_path_annotation.rbs @@ -11,7 +11,7 @@ module OpenAI type: :file_path } - class FilePathAnnotation < OpenAI::BaseModel + class FilePathAnnotation < OpenAI::Internal::Type::BaseModel attr_accessor end_index: Integer attr_accessor file_path: OpenAI::Models::Beta::Threads::FilePathAnnotation::FilePath @@ -34,7 +34,7 @@ module OpenAI type file_path = { file_id: String } - class FilePath < OpenAI::BaseModel + class FilePath < OpenAI::Internal::Type::BaseModel attr_accessor file_id: String def initialize: (file_id: String) -> void diff --git a/sig/openai/models/beta/threads/file_path_delta_annotation.rbs b/sig/openai/models/beta/threads/file_path_delta_annotation.rbs index 6c8e2380..671f758d 100644 --- a/sig/openai/models/beta/threads/file_path_delta_annotation.rbs +++ b/sig/openai/models/beta/threads/file_path_delta_annotation.rbs @@ -12,7 +12,7 @@ module OpenAI text: String } - class FilePathDeltaAnnotation < OpenAI::BaseModel + class FilePathDeltaAnnotation < OpenAI::Internal::Type::BaseModel attr_accessor index: Integer attr_accessor type: :file_path @@ -48,7 +48,7 @@ module OpenAI type file_path = { file_id: String } - class FilePath < OpenAI::BaseModel + class FilePath < OpenAI::Internal::Type::BaseModel attr_reader file_id: String? def file_id=: (String) -> String diff --git a/sig/openai/models/beta/threads/image_file.rbs b/sig/openai/models/beta/threads/image_file.rbs index 0fb507cc..8c594ae4 100644 --- a/sig/openai/models/beta/threads/image_file.rbs +++ b/sig/openai/models/beta/threads/image_file.rbs @@ -8,7 +8,7 @@ module OpenAI detail: OpenAI::Models::Beta::Threads::ImageFile::detail } - class ImageFile < OpenAI::BaseModel + class ImageFile < OpenAI::Internal::Type::BaseModel attr_accessor file_id: String attr_reader detail: OpenAI::Models::Beta::Threads::ImageFile::detail? @@ -27,7 +27,7 @@ module OpenAI type detail = :auto | :low | :high module Detail - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO: :auto LOW: :low diff --git a/sig/openai/models/beta/threads/image_file_content_block.rbs b/sig/openai/models/beta/threads/image_file_content_block.rbs index 7f0e99ee..85cbd424 100644 --- a/sig/openai/models/beta/threads/image_file_content_block.rbs +++ b/sig/openai/models/beta/threads/image_file_content_block.rbs @@ -8,7 +8,7 @@ module OpenAI type: :image_file } - class ImageFileContentBlock < OpenAI::BaseModel + class ImageFileContentBlock < OpenAI::Internal::Type::BaseModel attr_accessor image_file: OpenAI::Models::Beta::Threads::ImageFile attr_accessor type: :image_file diff --git a/sig/openai/models/beta/threads/image_file_delta.rbs b/sig/openai/models/beta/threads/image_file_delta.rbs index f645c00f..1360958e 100644 --- a/sig/openai/models/beta/threads/image_file_delta.rbs +++ b/sig/openai/models/beta/threads/image_file_delta.rbs @@ -8,7 +8,7 @@ module OpenAI file_id: String } - class ImageFileDelta < OpenAI::BaseModel + class ImageFileDelta < OpenAI::Internal::Type::BaseModel attr_reader detail: OpenAI::Models::Beta::Threads::ImageFileDelta::detail? def detail=: ( @@ -29,7 +29,7 @@ module OpenAI type detail = :auto | :low | :high module Detail - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO: :auto LOW: :low diff --git a/sig/openai/models/beta/threads/image_file_delta_block.rbs b/sig/openai/models/beta/threads/image_file_delta_block.rbs index e7e85e2b..cdb0e087 100644 --- a/sig/openai/models/beta/threads/image_file_delta_block.rbs +++ b/sig/openai/models/beta/threads/image_file_delta_block.rbs @@ -9,7 +9,7 @@ module OpenAI image_file: OpenAI::Models::Beta::Threads::ImageFileDelta } - class ImageFileDeltaBlock < OpenAI::BaseModel + class ImageFileDeltaBlock < OpenAI::Internal::Type::BaseModel attr_accessor index: Integer attr_accessor type: :image_file diff --git a/sig/openai/models/beta/threads/image_url.rbs b/sig/openai/models/beta/threads/image_url.rbs index 6039ca67..90adafe5 100644 --- a/sig/openai/models/beta/threads/image_url.rbs +++ b/sig/openai/models/beta/threads/image_url.rbs @@ -8,7 +8,7 @@ module OpenAI detail: OpenAI::Models::Beta::Threads::ImageURL::detail } - class ImageURL < OpenAI::BaseModel + class ImageURL < OpenAI::Internal::Type::BaseModel attr_accessor url: String attr_reader detail: OpenAI::Models::Beta::Threads::ImageURL::detail? @@ -27,7 +27,7 @@ module OpenAI type detail = :auto | :low | :high module Detail - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO: :auto LOW: :low diff --git a/sig/openai/models/beta/threads/image_url_content_block.rbs b/sig/openai/models/beta/threads/image_url_content_block.rbs index bd4eb8a7..a8e5425f 100644 --- a/sig/openai/models/beta/threads/image_url_content_block.rbs +++ b/sig/openai/models/beta/threads/image_url_content_block.rbs @@ -8,7 +8,7 @@ module OpenAI type: :image_url } - class ImageURLContentBlock < OpenAI::BaseModel + class ImageURLContentBlock < OpenAI::Internal::Type::BaseModel attr_accessor image_url: OpenAI::Models::Beta::Threads::ImageURL attr_accessor type: :image_url diff --git a/sig/openai/models/beta/threads/image_url_delta.rbs b/sig/openai/models/beta/threads/image_url_delta.rbs index 9cd1c721..b85ba1e9 100644 --- a/sig/openai/models/beta/threads/image_url_delta.rbs +++ b/sig/openai/models/beta/threads/image_url_delta.rbs @@ -8,7 +8,7 @@ module OpenAI url: String } - class ImageURLDelta < OpenAI::BaseModel + class ImageURLDelta < OpenAI::Internal::Type::BaseModel attr_reader detail: OpenAI::Models::Beta::Threads::ImageURLDelta::detail? def detail=: ( @@ -29,7 +29,7 @@ module OpenAI type detail = :auto | :low | :high module Detail - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO: :auto LOW: :low diff --git a/sig/openai/models/beta/threads/image_url_delta_block.rbs b/sig/openai/models/beta/threads/image_url_delta_block.rbs index 7e548334..4269c159 100644 --- a/sig/openai/models/beta/threads/image_url_delta_block.rbs +++ b/sig/openai/models/beta/threads/image_url_delta_block.rbs @@ -9,7 +9,7 @@ module OpenAI image_url: OpenAI::Models::Beta::Threads::ImageURLDelta } - class ImageURLDeltaBlock < OpenAI::BaseModel + class ImageURLDeltaBlock < OpenAI::Internal::Type::BaseModel attr_accessor index: Integer attr_accessor type: :image_url diff --git a/sig/openai/models/beta/threads/message.rbs b/sig/openai/models/beta/threads/message.rbs index e9a83633..02263056 100644 --- a/sig/openai/models/beta/threads/message.rbs +++ b/sig/openai/models/beta/threads/message.rbs @@ -20,7 +20,7 @@ module OpenAI thread_id: String } - class Message < OpenAI::BaseModel + class Message < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor assistant_id: String? @@ -74,7 +74,7 @@ module OpenAI tools: ::Array[OpenAI::Models::Beta::Threads::Message::Attachment::tool] } - class Attachment < OpenAI::BaseModel + class Attachment < OpenAI::Internal::Type::BaseModel attr_reader file_id: String? def file_id=: (String) -> String @@ -97,12 +97,12 @@ module OpenAI | OpenAI::Models::Beta::Threads::Message::Attachment::Tool::AssistantToolsFileSearchTypeOnly module Tool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union type assistant_tools_file_search_type_only = { type: :file_search } - class AssistantToolsFileSearchTypeOnly < OpenAI::BaseModel + class AssistantToolsFileSearchTypeOnly < OpenAI::Internal::Type::BaseModel attr_accessor type: :file_search def initialize: (?type: :file_search) -> void @@ -119,7 +119,7 @@ module OpenAI reason: OpenAI::Models::Beta::Threads::Message::IncompleteDetails::reason } - class IncompleteDetails < OpenAI::BaseModel + class IncompleteDetails < OpenAI::Internal::Type::BaseModel attr_accessor reason: OpenAI::Models::Beta::Threads::Message::IncompleteDetails::reason def initialize: ( @@ -136,7 +136,7 @@ module OpenAI | :run_failed module Reason - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum CONTENT_FILTER: :content_filter MAX_TOKENS: :max_tokens @@ -151,7 +151,7 @@ module OpenAI type role = :user | :assistant module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum USER: :user ASSISTANT: :assistant @@ -162,7 +162,7 @@ module OpenAI type status = :in_progress | :incomplete | :completed module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS: :in_progress INCOMPLETE: :incomplete diff --git a/sig/openai/models/beta/threads/message_content.rbs b/sig/openai/models/beta/threads/message_content.rbs index fda3bd19..e3bc2805 100644 --- a/sig/openai/models/beta/threads/message_content.rbs +++ b/sig/openai/models/beta/threads/message_content.rbs @@ -9,7 +9,7 @@ module OpenAI | OpenAI::Models::Beta::Threads::RefusalContentBlock module MessageContent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Beta::Threads::ImageFileContentBlock, OpenAI::Models::Beta::Threads::ImageURLContentBlock, OpenAI::Models::Beta::Threads::TextContentBlock, OpenAI::Models::Beta::Threads::RefusalContentBlock] end diff --git a/sig/openai/models/beta/threads/message_content_delta.rbs b/sig/openai/models/beta/threads/message_content_delta.rbs index c143f0ac..f06072ee 100644 --- a/sig/openai/models/beta/threads/message_content_delta.rbs +++ b/sig/openai/models/beta/threads/message_content_delta.rbs @@ -9,7 +9,7 @@ module OpenAI | OpenAI::Models::Beta::Threads::ImageURLDeltaBlock module MessageContentDelta - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Beta::Threads::ImageFileDeltaBlock, OpenAI::Models::Beta::Threads::TextDeltaBlock, OpenAI::Models::Beta::Threads::RefusalDeltaBlock, OpenAI::Models::Beta::Threads::ImageURLDeltaBlock] end diff --git a/sig/openai/models/beta/threads/message_content_part_param.rbs b/sig/openai/models/beta/threads/message_content_part_param.rbs index 9749af26..1e8a7d91 100644 --- a/sig/openai/models/beta/threads/message_content_part_param.rbs +++ b/sig/openai/models/beta/threads/message_content_part_param.rbs @@ -8,7 +8,7 @@ module OpenAI | OpenAI::Models::Beta::Threads::TextContentBlockParam module MessageContentPartParam - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Beta::Threads::ImageFileContentBlock, OpenAI::Models::Beta::Threads::ImageURLContentBlock, OpenAI::Models::Beta::Threads::TextContentBlockParam] end diff --git a/sig/openai/models/beta/threads/message_create_params.rbs b/sig/openai/models/beta/threads/message_create_params.rbs index 0091df98..0b4397bf 100644 --- a/sig/openai/models/beta/threads/message_create_params.rbs +++ b/sig/openai/models/beta/threads/message_create_params.rbs @@ -11,7 +11,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class MessageCreateParams < OpenAI::BaseModel + class MessageCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -38,7 +38,7 @@ module OpenAI | ::Array[OpenAI::Models::Beta::Threads::message_content_part_param] module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, ::Array[OpenAI::Models::Beta::Threads::message_content_part_param]] @@ -51,7 +51,7 @@ module OpenAI type role = :user | :assistant module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum USER: :user ASSISTANT: :assistant @@ -65,7 +65,7 @@ module OpenAI tools: ::Array[OpenAI::Models::Beta::Threads::MessageCreateParams::Attachment::tool] } - class Attachment < OpenAI::BaseModel + class Attachment < OpenAI::Internal::Type::BaseModel attr_reader file_id: String? def file_id=: (String) -> String @@ -88,11 +88,11 @@ module OpenAI | OpenAI::Models::Beta::Threads::MessageCreateParams::Attachment::Tool::FileSearch module Tool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union type file_search = { type: :file_search } - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel attr_accessor type: :file_search def initialize: (?type: :file_search) -> void diff --git a/sig/openai/models/beta/threads/message_delete_params.rbs b/sig/openai/models/beta/threads/message_delete_params.rbs index f400a51c..9abbe1c5 100644 --- a/sig/openai/models/beta/threads/message_delete_params.rbs +++ b/sig/openai/models/beta/threads/message_delete_params.rbs @@ -5,7 +5,7 @@ module OpenAI type message_delete_params = { thread_id: String } & OpenAI::Internal::Type::request_parameters - class MessageDeleteParams < OpenAI::BaseModel + class MessageDeleteParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/beta/threads/message_deleted.rbs b/sig/openai/models/beta/threads/message_deleted.rbs index f6d3a96e..d1356267 100644 --- a/sig/openai/models/beta/threads/message_deleted.rbs +++ b/sig/openai/models/beta/threads/message_deleted.rbs @@ -5,7 +5,7 @@ module OpenAI type message_deleted = { id: String, deleted: bool, object: :"thread.message.deleted" } - class MessageDeleted < OpenAI::BaseModel + class MessageDeleted < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor deleted: bool diff --git a/sig/openai/models/beta/threads/message_delta.rbs b/sig/openai/models/beta/threads/message_delta.rbs index f40f44ea..e74b69e1 100644 --- a/sig/openai/models/beta/threads/message_delta.rbs +++ b/sig/openai/models/beta/threads/message_delta.rbs @@ -8,7 +8,7 @@ module OpenAI role: OpenAI::Models::Beta::Threads::MessageDelta::role } - class MessageDelta < OpenAI::BaseModel + class MessageDelta < OpenAI::Internal::Type::BaseModel attr_reader content: ::Array[OpenAI::Models::Beta::Threads::message_content_delta]? def content=: ( @@ -31,7 +31,7 @@ module OpenAI type role = :user | :assistant module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum USER: :user ASSISTANT: :assistant diff --git a/sig/openai/models/beta/threads/message_delta_event.rbs b/sig/openai/models/beta/threads/message_delta_event.rbs index d34c13f6..553d6fef 100644 --- a/sig/openai/models/beta/threads/message_delta_event.rbs +++ b/sig/openai/models/beta/threads/message_delta_event.rbs @@ -9,7 +9,7 @@ module OpenAI object: :"thread.message.delta" } - class MessageDeltaEvent < OpenAI::BaseModel + class MessageDeltaEvent < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor delta: OpenAI::Models::Beta::Threads::MessageDelta diff --git a/sig/openai/models/beta/threads/message_list_params.rbs b/sig/openai/models/beta/threads/message_list_params.rbs index 9b331b46..2cdda317 100644 --- a/sig/openai/models/beta/threads/message_list_params.rbs +++ b/sig/openai/models/beta/threads/message_list_params.rbs @@ -12,7 +12,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class MessageListParams < OpenAI::BaseModel + class MessageListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -52,7 +52,7 @@ module OpenAI type order = :asc | :desc module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASC: :asc DESC: :desc diff --git a/sig/openai/models/beta/threads/message_retrieve_params.rbs b/sig/openai/models/beta/threads/message_retrieve_params.rbs index 58434310..c6ff2a71 100644 --- a/sig/openai/models/beta/threads/message_retrieve_params.rbs +++ b/sig/openai/models/beta/threads/message_retrieve_params.rbs @@ -5,7 +5,7 @@ module OpenAI type message_retrieve_params = { thread_id: String } & OpenAI::Internal::Type::request_parameters - class MessageRetrieveParams < OpenAI::BaseModel + class MessageRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/beta/threads/message_update_params.rbs b/sig/openai/models/beta/threads/message_update_params.rbs index 7d14cb66..4f14f212 100644 --- a/sig/openai/models/beta/threads/message_update_params.rbs +++ b/sig/openai/models/beta/threads/message_update_params.rbs @@ -6,7 +6,7 @@ module OpenAI { thread_id: String, metadata: OpenAI::Models::metadata? } & OpenAI::Internal::Type::request_parameters - class MessageUpdateParams < OpenAI::BaseModel + class MessageUpdateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/beta/threads/refusal_content_block.rbs b/sig/openai/models/beta/threads/refusal_content_block.rbs index 041435a5..c91e1603 100644 --- a/sig/openai/models/beta/threads/refusal_content_block.rbs +++ b/sig/openai/models/beta/threads/refusal_content_block.rbs @@ -4,7 +4,7 @@ module OpenAI module Threads type refusal_content_block = { refusal: String, type: :refusal } - class RefusalContentBlock < OpenAI::BaseModel + class RefusalContentBlock < OpenAI::Internal::Type::BaseModel attr_accessor refusal: String attr_accessor type: :refusal diff --git a/sig/openai/models/beta/threads/refusal_delta_block.rbs b/sig/openai/models/beta/threads/refusal_delta_block.rbs index 635407d5..6fc97486 100644 --- a/sig/openai/models/beta/threads/refusal_delta_block.rbs +++ b/sig/openai/models/beta/threads/refusal_delta_block.rbs @@ -5,7 +5,7 @@ module OpenAI type refusal_delta_block = { index: Integer, type: :refusal, refusal: String } - class RefusalDeltaBlock < OpenAI::BaseModel + class RefusalDeltaBlock < OpenAI::Internal::Type::BaseModel attr_accessor index: Integer attr_accessor type: :refusal diff --git a/sig/openai/models/beta/threads/required_action_function_tool_call.rbs b/sig/openai/models/beta/threads/required_action_function_tool_call.rbs index 9429819f..b2cd22a2 100644 --- a/sig/openai/models/beta/threads/required_action_function_tool_call.rbs +++ b/sig/openai/models/beta/threads/required_action_function_tool_call.rbs @@ -9,7 +9,7 @@ module OpenAI type: :function } - class RequiredActionFunctionToolCall < OpenAI::BaseModel + class RequiredActionFunctionToolCall < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor function: OpenAI::Models::Beta::Threads::RequiredActionFunctionToolCall::Function @@ -26,7 +26,7 @@ module OpenAI type function = { arguments: String, name: String } - class Function < OpenAI::BaseModel + class Function < OpenAI::Internal::Type::BaseModel attr_accessor arguments: String attr_accessor name: String diff --git a/sig/openai/models/beta/threads/run.rbs b/sig/openai/models/beta/threads/run.rbs index c36d145c..6fb8ca5b 100644 --- a/sig/openai/models/beta/threads/run.rbs +++ b/sig/openai/models/beta/threads/run.rbs @@ -33,7 +33,7 @@ module OpenAI top_p: Float? } - class Run < OpenAI::BaseModel + class Run < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor assistant_id: String @@ -125,7 +125,7 @@ module OpenAI reason: OpenAI::Models::Beta::Threads::Run::IncompleteDetails::reason } - class IncompleteDetails < OpenAI::BaseModel + class IncompleteDetails < OpenAI::Internal::Type::BaseModel attr_reader reason: OpenAI::Models::Beta::Threads::Run::IncompleteDetails::reason? def reason=: ( @@ -141,7 +141,7 @@ module OpenAI type reason = :max_completion_tokens | :max_prompt_tokens module Reason - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum MAX_COMPLETION_TOKENS: :max_completion_tokens MAX_PROMPT_TOKENS: :max_prompt_tokens @@ -156,7 +156,7 @@ module OpenAI message: String } - class LastError < OpenAI::BaseModel + class LastError < OpenAI::Internal::Type::BaseModel attr_accessor code: OpenAI::Models::Beta::Threads::Run::LastError::code attr_accessor message: String @@ -171,7 +171,7 @@ module OpenAI type code = :server_error | :rate_limit_exceeded | :invalid_prompt module Code - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum SERVER_ERROR: :server_error RATE_LIMIT_EXCEEDED: :rate_limit_exceeded @@ -187,7 +187,7 @@ module OpenAI type: :submit_tool_outputs } - class RequiredAction < OpenAI::BaseModel + class RequiredAction < OpenAI::Internal::Type::BaseModel attr_accessor submit_tool_outputs: OpenAI::Models::Beta::Threads::Run::RequiredAction::SubmitToolOutputs attr_accessor type: :submit_tool_outputs @@ -204,7 +204,7 @@ module OpenAI tool_calls: ::Array[OpenAI::Models::Beta::Threads::RequiredActionFunctionToolCall] } - class SubmitToolOutputs < OpenAI::BaseModel + class SubmitToolOutputs < OpenAI::Internal::Type::BaseModel attr_accessor tool_calls: ::Array[OpenAI::Models::Beta::Threads::RequiredActionFunctionToolCall] def initialize: ( @@ -221,7 +221,7 @@ module OpenAI last_messages: Integer? } - class TruncationStrategy < OpenAI::BaseModel + class TruncationStrategy < OpenAI::Internal::Type::BaseModel attr_accessor type: OpenAI::Models::Beta::Threads::Run::TruncationStrategy::type_ attr_accessor last_messages: Integer? @@ -236,7 +236,7 @@ module OpenAI type type_ = :auto | :last_messages module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO: :auto LAST_MESSAGES: :last_messages @@ -252,7 +252,7 @@ module OpenAI total_tokens: Integer } - class Usage < OpenAI::BaseModel + class Usage < OpenAI::Internal::Type::BaseModel attr_accessor completion_tokens: Integer attr_accessor prompt_tokens: Integer diff --git a/sig/openai/models/beta/threads/run_cancel_params.rbs b/sig/openai/models/beta/threads/run_cancel_params.rbs index 5460b5ad..4c384a32 100644 --- a/sig/openai/models/beta/threads/run_cancel_params.rbs +++ b/sig/openai/models/beta/threads/run_cancel_params.rbs @@ -5,7 +5,7 @@ module OpenAI type run_cancel_params = { thread_id: String } & OpenAI::Internal::Type::request_parameters - class RunCancelParams < OpenAI::BaseModel + class RunCancelParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/beta/threads/run_create_params.rbs b/sig/openai/models/beta/threads/run_create_params.rbs index 360db47b..f5462f0c 100644 --- a/sig/openai/models/beta/threads/run_create_params.rbs +++ b/sig/openai/models/beta/threads/run_create_params.rbs @@ -24,7 +24,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class RunCreateParams < OpenAI::BaseModel + class RunCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -99,7 +99,7 @@ module OpenAI metadata: OpenAI::Models::metadata? } - class AdditionalMessage < OpenAI::BaseModel + class AdditionalMessage < OpenAI::Internal::Type::BaseModel attr_accessor content: OpenAI::Models::Beta::Threads::RunCreateParams::AdditionalMessage::content attr_accessor role: OpenAI::Models::Beta::Threads::RunCreateParams::AdditionalMessage::role @@ -122,7 +122,7 @@ module OpenAI | ::Array[OpenAI::Models::Beta::Threads::message_content_part_param] module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, ::Array[OpenAI::Models::Beta::Threads::message_content_part_param]] @@ -135,7 +135,7 @@ module OpenAI type role = :user | :assistant module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum USER: :user ASSISTANT: :assistant @@ -149,7 +149,7 @@ module OpenAI tools: ::Array[OpenAI::Models::Beta::Threads::RunCreateParams::AdditionalMessage::Attachment::tool] } - class Attachment < OpenAI::BaseModel + class Attachment < OpenAI::Internal::Type::BaseModel attr_reader file_id: String? def file_id=: (String) -> String @@ -172,11 +172,11 @@ module OpenAI | OpenAI::Models::Beta::Threads::RunCreateParams::AdditionalMessage::Attachment::Tool::FileSearch module Tool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union type file_search = { type: :file_search } - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel attr_accessor type: :file_search def initialize: (?type: :file_search) -> void @@ -192,7 +192,7 @@ module OpenAI type model = String | OpenAI::Models::chat_model module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, OpenAI::Models::chat_model] end @@ -203,7 +203,7 @@ module OpenAI last_messages: Integer? } - class TruncationStrategy < OpenAI::BaseModel + class TruncationStrategy < OpenAI::Internal::Type::BaseModel attr_accessor type: OpenAI::Models::Beta::Threads::RunCreateParams::TruncationStrategy::type_ attr_accessor last_messages: Integer? @@ -218,7 +218,7 @@ module OpenAI type type_ = :auto | :last_messages module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO: :auto LAST_MESSAGES: :last_messages diff --git a/sig/openai/models/beta/threads/run_list_params.rbs b/sig/openai/models/beta/threads/run_list_params.rbs index 525ad66f..42825950 100644 --- a/sig/openai/models/beta/threads/run_list_params.rbs +++ b/sig/openai/models/beta/threads/run_list_params.rbs @@ -11,7 +11,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class RunListParams < OpenAI::BaseModel + class RunListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -46,7 +46,7 @@ module OpenAI type order = :asc | :desc module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASC: :asc DESC: :desc diff --git a/sig/openai/models/beta/threads/run_retrieve_params.rbs b/sig/openai/models/beta/threads/run_retrieve_params.rbs index a76a9131..ddbeaf2b 100644 --- a/sig/openai/models/beta/threads/run_retrieve_params.rbs +++ b/sig/openai/models/beta/threads/run_retrieve_params.rbs @@ -5,7 +5,7 @@ module OpenAI type run_retrieve_params = { thread_id: String } & OpenAI::Internal::Type::request_parameters - class RunRetrieveParams < OpenAI::BaseModel + class RunRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/beta/threads/run_status.rbs b/sig/openai/models/beta/threads/run_status.rbs index 4b0e430b..4a106ac0 100644 --- a/sig/openai/models/beta/threads/run_status.rbs +++ b/sig/openai/models/beta/threads/run_status.rbs @@ -14,7 +14,7 @@ module OpenAI | :expired module RunStatus - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum QUEUED: :queued IN_PROGRESS: :in_progress diff --git a/sig/openai/models/beta/threads/run_submit_tool_outputs_params.rbs b/sig/openai/models/beta/threads/run_submit_tool_outputs_params.rbs index 48e05b57..294478bc 100644 --- a/sig/openai/models/beta/threads/run_submit_tool_outputs_params.rbs +++ b/sig/openai/models/beta/threads/run_submit_tool_outputs_params.rbs @@ -9,7 +9,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class RunSubmitToolOutputsParams < OpenAI::BaseModel + class RunSubmitToolOutputsParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -27,7 +27,7 @@ module OpenAI type tool_output = { output: String, tool_call_id: String } - class ToolOutput < OpenAI::BaseModel + class ToolOutput < OpenAI::Internal::Type::BaseModel attr_reader output: String? def output=: (String) -> String diff --git a/sig/openai/models/beta/threads/run_update_params.rbs b/sig/openai/models/beta/threads/run_update_params.rbs index 6e46178a..8ea338f3 100644 --- a/sig/openai/models/beta/threads/run_update_params.rbs +++ b/sig/openai/models/beta/threads/run_update_params.rbs @@ -6,7 +6,7 @@ module OpenAI { thread_id: String, metadata: OpenAI::Models::metadata? } & OpenAI::Internal::Type::request_parameters - class RunUpdateParams < OpenAI::BaseModel + class RunUpdateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/beta/threads/runs/code_interpreter_logs.rbs b/sig/openai/models/beta/threads/runs/code_interpreter_logs.rbs index 85c320b0..e002868b 100644 --- a/sig/openai/models/beta/threads/runs/code_interpreter_logs.rbs +++ b/sig/openai/models/beta/threads/runs/code_interpreter_logs.rbs @@ -6,7 +6,7 @@ module OpenAI type code_interpreter_logs = { index: Integer, type: :logs, logs: String } - class CodeInterpreterLogs < OpenAI::BaseModel + class CodeInterpreterLogs < OpenAI::Internal::Type::BaseModel attr_accessor index: Integer attr_accessor type: :logs diff --git a/sig/openai/models/beta/threads/runs/code_interpreter_output_image.rbs b/sig/openai/models/beta/threads/runs/code_interpreter_output_image.rbs index f640b74e..e05b9326 100644 --- a/sig/openai/models/beta/threads/runs/code_interpreter_output_image.rbs +++ b/sig/openai/models/beta/threads/runs/code_interpreter_output_image.rbs @@ -10,7 +10,7 @@ module OpenAI image: OpenAI::Models::Beta::Threads::Runs::CodeInterpreterOutputImage::Image } - class CodeInterpreterOutputImage < OpenAI::BaseModel + class CodeInterpreterOutputImage < OpenAI::Internal::Type::BaseModel attr_accessor index: Integer attr_accessor type: :image @@ -31,7 +31,7 @@ module OpenAI type image = { file_id: String } - class Image < OpenAI::BaseModel + class Image < OpenAI::Internal::Type::BaseModel attr_reader file_id: String? def file_id=: (String) -> String diff --git a/sig/openai/models/beta/threads/runs/code_interpreter_tool_call.rbs b/sig/openai/models/beta/threads/runs/code_interpreter_tool_call.rbs index fa6f7865..6ded6a40 100644 --- a/sig/openai/models/beta/threads/runs/code_interpreter_tool_call.rbs +++ b/sig/openai/models/beta/threads/runs/code_interpreter_tool_call.rbs @@ -10,7 +10,7 @@ module OpenAI type: :code_interpreter } - class CodeInterpreterToolCall < OpenAI::BaseModel + class CodeInterpreterToolCall < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor code_interpreter: OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCall::CodeInterpreter @@ -31,7 +31,7 @@ module OpenAI outputs: ::Array[OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCall::CodeInterpreter::output] } - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel attr_accessor input: String attr_accessor outputs: ::Array[OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCall::CodeInterpreter::output] @@ -48,11 +48,11 @@ module OpenAI | OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCall::CodeInterpreter::Output::Image module Output - extend OpenAI::Union + extend OpenAI::Internal::Type::Union type logs = { logs: String, type: :logs } - class Logs < OpenAI::BaseModel + class Logs < OpenAI::Internal::Type::BaseModel attr_accessor logs: String attr_accessor type: :logs @@ -68,7 +68,7 @@ module OpenAI type: :image } - class Image < OpenAI::BaseModel + class Image < OpenAI::Internal::Type::BaseModel attr_accessor image: OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCall::CodeInterpreter::Output::Image::Image attr_accessor type: :image @@ -82,7 +82,7 @@ module OpenAI type image = { file_id: String } - class Image < OpenAI::BaseModel + class Image < OpenAI::Internal::Type::BaseModel attr_accessor file_id: String def initialize: (file_id: String) -> void diff --git a/sig/openai/models/beta/threads/runs/code_interpreter_tool_call_delta.rbs b/sig/openai/models/beta/threads/runs/code_interpreter_tool_call_delta.rbs index 32dc3fb8..632bed4c 100644 --- a/sig/openai/models/beta/threads/runs/code_interpreter_tool_call_delta.rbs +++ b/sig/openai/models/beta/threads/runs/code_interpreter_tool_call_delta.rbs @@ -11,7 +11,7 @@ module OpenAI code_interpreter: OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCallDelta::CodeInterpreter } - class CodeInterpreterToolCallDelta < OpenAI::BaseModel + class CodeInterpreterToolCallDelta < OpenAI::Internal::Type::BaseModel attr_accessor index: Integer attr_accessor type: :code_interpreter @@ -41,7 +41,7 @@ module OpenAI outputs: ::Array[OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCallDelta::CodeInterpreter::output] } - class CodeInterpreter < OpenAI::BaseModel + class CodeInterpreter < OpenAI::Internal::Type::BaseModel attr_reader input: String? def input=: (String) -> String @@ -64,7 +64,7 @@ module OpenAI | OpenAI::Models::Beta::Threads::Runs::CodeInterpreterOutputImage module Output - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Beta::Threads::Runs::CodeInterpreterLogs, OpenAI::Models::Beta::Threads::Runs::CodeInterpreterOutputImage] end diff --git a/sig/openai/models/beta/threads/runs/file_search_tool_call.rbs b/sig/openai/models/beta/threads/runs/file_search_tool_call.rbs index f5a993d3..c0437f96 100644 --- a/sig/openai/models/beta/threads/runs/file_search_tool_call.rbs +++ b/sig/openai/models/beta/threads/runs/file_search_tool_call.rbs @@ -10,7 +10,7 @@ module OpenAI type: :file_search } - class FileSearchToolCall < OpenAI::BaseModel + class FileSearchToolCall < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor file_search: OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch @@ -31,7 +31,7 @@ module OpenAI results: ::Array[OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch::Result] } - class FileSearch < OpenAI::BaseModel + class FileSearch < OpenAI::Internal::Type::BaseModel attr_reader ranking_options: OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch::RankingOptions? def ranking_options=: ( @@ -57,7 +57,7 @@ module OpenAI score_threshold: Float } - class RankingOptions < OpenAI::BaseModel + class RankingOptions < OpenAI::Internal::Type::BaseModel attr_accessor ranker: OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch::RankingOptions::ranker attr_accessor score_threshold: Float @@ -72,7 +72,7 @@ module OpenAI type ranker = :auto | :default_2024_08_21 module Ranker - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO: :auto DEFAULT_2024_08_21: :default_2024_08_21 @@ -89,7 +89,7 @@ module OpenAI content: ::Array[OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch::Result::Content] } - class Result < OpenAI::BaseModel + class Result < OpenAI::Internal::Type::BaseModel attr_accessor file_id: String attr_accessor file_name: String @@ -117,7 +117,7 @@ module OpenAI type: OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall::FileSearch::Result::Content::type_ } - class Content < OpenAI::BaseModel + class Content < OpenAI::Internal::Type::BaseModel attr_reader text: String? def text=: (String) -> String @@ -138,7 +138,7 @@ module OpenAI type type_ = :text module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT: :text diff --git a/sig/openai/models/beta/threads/runs/file_search_tool_call_delta.rbs b/sig/openai/models/beta/threads/runs/file_search_tool_call_delta.rbs index 11ae3dc7..563a191d 100644 --- a/sig/openai/models/beta/threads/runs/file_search_tool_call_delta.rbs +++ b/sig/openai/models/beta/threads/runs/file_search_tool_call_delta.rbs @@ -6,7 +6,7 @@ module OpenAI type file_search_tool_call_delta = { file_search: top, index: Integer, type: :file_search, id: String } - class FileSearchToolCallDelta < OpenAI::BaseModel + class FileSearchToolCallDelta < OpenAI::Internal::Type::BaseModel attr_accessor file_search: top attr_accessor index: Integer diff --git a/sig/openai/models/beta/threads/runs/function_tool_call.rbs b/sig/openai/models/beta/threads/runs/function_tool_call.rbs index c3e9168a..324286aa 100644 --- a/sig/openai/models/beta/threads/runs/function_tool_call.rbs +++ b/sig/openai/models/beta/threads/runs/function_tool_call.rbs @@ -10,7 +10,7 @@ module OpenAI type: :function } - class FunctionToolCall < OpenAI::BaseModel + class FunctionToolCall < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor function: OpenAI::Models::Beta::Threads::Runs::FunctionToolCall::Function @@ -27,7 +27,7 @@ module OpenAI type function = { arguments: String, name: String, output: String? } - class Function < OpenAI::BaseModel + class Function < OpenAI::Internal::Type::BaseModel attr_accessor arguments: String attr_accessor name: String diff --git a/sig/openai/models/beta/threads/runs/function_tool_call_delta.rbs b/sig/openai/models/beta/threads/runs/function_tool_call_delta.rbs index f6a01247..0f96b130 100644 --- a/sig/openai/models/beta/threads/runs/function_tool_call_delta.rbs +++ b/sig/openai/models/beta/threads/runs/function_tool_call_delta.rbs @@ -11,7 +11,7 @@ module OpenAI function: OpenAI::Models::Beta::Threads::Runs::FunctionToolCallDelta::Function } - class FunctionToolCallDelta < OpenAI::BaseModel + class FunctionToolCallDelta < OpenAI::Internal::Type::BaseModel attr_accessor index: Integer attr_accessor type: :function @@ -37,7 +37,7 @@ module OpenAI type function = { arguments: String, name: String, output: String? } - class Function < OpenAI::BaseModel + class Function < OpenAI::Internal::Type::BaseModel attr_reader arguments: String? def arguments=: (String) -> String diff --git a/sig/openai/models/beta/threads/runs/message_creation_step_details.rbs b/sig/openai/models/beta/threads/runs/message_creation_step_details.rbs index 3ddacf7b..f7300f72 100644 --- a/sig/openai/models/beta/threads/runs/message_creation_step_details.rbs +++ b/sig/openai/models/beta/threads/runs/message_creation_step_details.rbs @@ -9,7 +9,7 @@ module OpenAI type: :message_creation } - class MessageCreationStepDetails < OpenAI::BaseModel + class MessageCreationStepDetails < OpenAI::Internal::Type::BaseModel attr_accessor message_creation: OpenAI::Models::Beta::Threads::Runs::MessageCreationStepDetails::MessageCreation attr_accessor type: :message_creation @@ -23,7 +23,7 @@ module OpenAI type message_creation = { message_id: String } - class MessageCreation < OpenAI::BaseModel + class MessageCreation < OpenAI::Internal::Type::BaseModel attr_accessor message_id: String def initialize: (message_id: String) -> void diff --git a/sig/openai/models/beta/threads/runs/run_step.rbs b/sig/openai/models/beta/threads/runs/run_step.rbs index a60e10a6..952766af 100644 --- a/sig/openai/models/beta/threads/runs/run_step.rbs +++ b/sig/openai/models/beta/threads/runs/run_step.rbs @@ -25,7 +25,7 @@ module OpenAI usage: OpenAI::Models::Beta::Threads::Runs::RunStep::Usage? } - class RunStep < OpenAI::BaseModel + class RunStep < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor assistant_id: String @@ -85,7 +85,7 @@ module OpenAI message: String } - class LastError < OpenAI::BaseModel + class LastError < OpenAI::Internal::Type::BaseModel attr_accessor code: OpenAI::Models::Beta::Threads::Runs::RunStep::LastError::code attr_accessor message: String @@ -100,7 +100,7 @@ module OpenAI type code = :server_error | :rate_limit_exceeded module Code - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum SERVER_ERROR: :server_error RATE_LIMIT_EXCEEDED: :rate_limit_exceeded @@ -113,7 +113,7 @@ module OpenAI :in_progress | :cancelled | :failed | :completed | :expired module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS: :in_progress CANCELLED: :cancelled @@ -129,7 +129,7 @@ module OpenAI | OpenAI::Models::Beta::Threads::Runs::ToolCallsStepDetails module StepDetails - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Beta::Threads::Runs::MessageCreationStepDetails, OpenAI::Models::Beta::Threads::Runs::ToolCallsStepDetails] end @@ -137,7 +137,7 @@ module OpenAI type type_ = :message_creation | :tool_calls module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum MESSAGE_CREATION: :message_creation TOOL_CALLS: :tool_calls @@ -152,7 +152,7 @@ module OpenAI total_tokens: Integer } - class Usage < OpenAI::BaseModel + class Usage < OpenAI::Internal::Type::BaseModel attr_accessor completion_tokens: Integer attr_accessor prompt_tokens: Integer diff --git a/sig/openai/models/beta/threads/runs/run_step_delta.rbs b/sig/openai/models/beta/threads/runs/run_step_delta.rbs index 45b48449..1f3143ee 100644 --- a/sig/openai/models/beta/threads/runs/run_step_delta.rbs +++ b/sig/openai/models/beta/threads/runs/run_step_delta.rbs @@ -10,7 +10,7 @@ module OpenAI step_details: OpenAI::Models::Beta::Threads::Runs::RunStepDelta::step_details } - class RunStepDelta < OpenAI::BaseModel + class RunStepDelta < OpenAI::Internal::Type::BaseModel attr_reader step_details: OpenAI::Models::Beta::Threads::Runs::RunStepDelta::step_details? def step_details=: ( @@ -28,7 +28,7 @@ module OpenAI | OpenAI::Models::Beta::Threads::Runs::ToolCallDeltaObject module StepDetails - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Beta::Threads::Runs::RunStepDeltaMessageDelta, OpenAI::Models::Beta::Threads::Runs::ToolCallDeltaObject] end diff --git a/sig/openai/models/beta/threads/runs/run_step_delta_event.rbs b/sig/openai/models/beta/threads/runs/run_step_delta_event.rbs index 295594a8..0300ea33 100644 --- a/sig/openai/models/beta/threads/runs/run_step_delta_event.rbs +++ b/sig/openai/models/beta/threads/runs/run_step_delta_event.rbs @@ -12,7 +12,7 @@ module OpenAI object: :"thread.run.step.delta" } - class RunStepDeltaEvent < OpenAI::BaseModel + class RunStepDeltaEvent < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor delta: OpenAI::Models::Beta::Threads::Runs::RunStepDelta diff --git a/sig/openai/models/beta/threads/runs/run_step_delta_message_delta.rbs b/sig/openai/models/beta/threads/runs/run_step_delta_message_delta.rbs index d1daf15e..b397b858 100644 --- a/sig/openai/models/beta/threads/runs/run_step_delta_message_delta.rbs +++ b/sig/openai/models/beta/threads/runs/run_step_delta_message_delta.rbs @@ -11,7 +11,7 @@ module OpenAI message_creation: OpenAI::Models::Beta::Threads::Runs::RunStepDeltaMessageDelta::MessageCreation } - class RunStepDeltaMessageDelta < OpenAI::BaseModel + class RunStepDeltaMessageDelta < OpenAI::Internal::Type::BaseModel attr_accessor type: :message_creation attr_reader message_creation: OpenAI::Models::Beta::Threads::Runs::RunStepDeltaMessageDelta::MessageCreation? @@ -29,7 +29,7 @@ module OpenAI type message_creation = { message_id: String } - class MessageCreation < OpenAI::BaseModel + class MessageCreation < OpenAI::Internal::Type::BaseModel attr_reader message_id: String? def message_id=: (String) -> String diff --git a/sig/openai/models/beta/threads/runs/run_step_include.rbs b/sig/openai/models/beta/threads/runs/run_step_include.rbs index 529038db..30d9eb94 100644 --- a/sig/openai/models/beta/threads/runs/run_step_include.rbs +++ b/sig/openai/models/beta/threads/runs/run_step_include.rbs @@ -9,7 +9,7 @@ module OpenAI :"step_details.tool_calls[*].file_search.results[*].content" module RunStepInclude - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum STEP_DETAILS_TOOL_CALLS_FILE_SEARCH_RESULTS_CONTENT: :"step_details.tool_calls[*].file_search.results[*].content" diff --git a/sig/openai/models/beta/threads/runs/step_list_params.rbs b/sig/openai/models/beta/threads/runs/step_list_params.rbs index 04b7cf53..5fe683ff 100644 --- a/sig/openai/models/beta/threads/runs/step_list_params.rbs +++ b/sig/openai/models/beta/threads/runs/step_list_params.rbs @@ -14,7 +14,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class StepListParams < OpenAI::BaseModel + class StepListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -59,7 +59,7 @@ module OpenAI type order = :asc | :desc module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASC: :asc DESC: :desc diff --git a/sig/openai/models/beta/threads/runs/step_retrieve_params.rbs b/sig/openai/models/beta/threads/runs/step_retrieve_params.rbs index 06b14341..c759864c 100644 --- a/sig/openai/models/beta/threads/runs/step_retrieve_params.rbs +++ b/sig/openai/models/beta/threads/runs/step_retrieve_params.rbs @@ -11,7 +11,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class StepRetrieveParams < OpenAI::BaseModel + class StepRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/beta/threads/runs/tool_call.rbs b/sig/openai/models/beta/threads/runs/tool_call.rbs index 2d13d06d..f3604833 100644 --- a/sig/openai/models/beta/threads/runs/tool_call.rbs +++ b/sig/openai/models/beta/threads/runs/tool_call.rbs @@ -9,7 +9,7 @@ module OpenAI | OpenAI::Models::Beta::Threads::Runs::FunctionToolCall module ToolCall - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCall, OpenAI::Models::Beta::Threads::Runs::FileSearchToolCall, OpenAI::Models::Beta::Threads::Runs::FunctionToolCall] end diff --git a/sig/openai/models/beta/threads/runs/tool_call_delta.rbs b/sig/openai/models/beta/threads/runs/tool_call_delta.rbs index 2e349dcd..f5159374 100644 --- a/sig/openai/models/beta/threads/runs/tool_call_delta.rbs +++ b/sig/openai/models/beta/threads/runs/tool_call_delta.rbs @@ -9,7 +9,7 @@ module OpenAI | OpenAI::Models::Beta::Threads::Runs::FunctionToolCallDelta module ToolCallDelta - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Beta::Threads::Runs::CodeInterpreterToolCallDelta, OpenAI::Models::Beta::Threads::Runs::FileSearchToolCallDelta, OpenAI::Models::Beta::Threads::Runs::FunctionToolCallDelta] end diff --git a/sig/openai/models/beta/threads/runs/tool_call_delta_object.rbs b/sig/openai/models/beta/threads/runs/tool_call_delta_object.rbs index 50619aae..f693e116 100644 --- a/sig/openai/models/beta/threads/runs/tool_call_delta_object.rbs +++ b/sig/openai/models/beta/threads/runs/tool_call_delta_object.rbs @@ -9,7 +9,7 @@ module OpenAI tool_calls: ::Array[OpenAI::Models::Beta::Threads::Runs::tool_call_delta] } - class ToolCallDeltaObject < OpenAI::BaseModel + class ToolCallDeltaObject < OpenAI::Internal::Type::BaseModel attr_accessor type: :tool_calls attr_reader tool_calls: ::Array[OpenAI::Models::Beta::Threads::Runs::tool_call_delta]? diff --git a/sig/openai/models/beta/threads/runs/tool_calls_step_details.rbs b/sig/openai/models/beta/threads/runs/tool_calls_step_details.rbs index 7f5c8851..c08bece5 100644 --- a/sig/openai/models/beta/threads/runs/tool_calls_step_details.rbs +++ b/sig/openai/models/beta/threads/runs/tool_calls_step_details.rbs @@ -9,7 +9,7 @@ module OpenAI type: :tool_calls } - class ToolCallsStepDetails < OpenAI::BaseModel + class ToolCallsStepDetails < OpenAI::Internal::Type::BaseModel attr_accessor tool_calls: ::Array[OpenAI::Models::Beta::Threads::Runs::tool_call] attr_accessor type: :tool_calls diff --git a/sig/openai/models/beta/threads/text.rbs b/sig/openai/models/beta/threads/text.rbs index 1c4ebeb3..a4aa3141 100644 --- a/sig/openai/models/beta/threads/text.rbs +++ b/sig/openai/models/beta/threads/text.rbs @@ -8,7 +8,7 @@ module OpenAI value: String } - class Text < OpenAI::BaseModel + class Text < OpenAI::Internal::Type::BaseModel attr_accessor annotations: ::Array[OpenAI::Models::Beta::Threads::annotation] attr_accessor value: String diff --git a/sig/openai/models/beta/threads/text_content_block.rbs b/sig/openai/models/beta/threads/text_content_block.rbs index b83f4585..2bcf7f2a 100644 --- a/sig/openai/models/beta/threads/text_content_block.rbs +++ b/sig/openai/models/beta/threads/text_content_block.rbs @@ -5,7 +5,7 @@ module OpenAI type text_content_block = { text: OpenAI::Models::Beta::Threads::Text, type: :text } - class TextContentBlock < OpenAI::BaseModel + class TextContentBlock < OpenAI::Internal::Type::BaseModel attr_accessor text: OpenAI::Models::Beta::Threads::Text attr_accessor type: :text diff --git a/sig/openai/models/beta/threads/text_content_block_param.rbs b/sig/openai/models/beta/threads/text_content_block_param.rbs index bf0b5b7e..291be17d 100644 --- a/sig/openai/models/beta/threads/text_content_block_param.rbs +++ b/sig/openai/models/beta/threads/text_content_block_param.rbs @@ -4,7 +4,7 @@ module OpenAI module Threads type text_content_block_param = { text: String, type: :text } - class TextContentBlockParam < OpenAI::BaseModel + class TextContentBlockParam < OpenAI::Internal::Type::BaseModel attr_accessor text: String attr_accessor type: :text diff --git a/sig/openai/models/beta/threads/text_delta.rbs b/sig/openai/models/beta/threads/text_delta.rbs index 721b012e..0b5ac945 100644 --- a/sig/openai/models/beta/threads/text_delta.rbs +++ b/sig/openai/models/beta/threads/text_delta.rbs @@ -8,7 +8,7 @@ module OpenAI value: String } - class TextDelta < OpenAI::BaseModel + class TextDelta < OpenAI::Internal::Type::BaseModel attr_reader annotations: ::Array[OpenAI::Models::Beta::Threads::annotation_delta]? def annotations=: ( diff --git a/sig/openai/models/beta/threads/text_delta_block.rbs b/sig/openai/models/beta/threads/text_delta_block.rbs index 2a03d985..8d70d40b 100644 --- a/sig/openai/models/beta/threads/text_delta_block.rbs +++ b/sig/openai/models/beta/threads/text_delta_block.rbs @@ -9,7 +9,7 @@ module OpenAI text: OpenAI::Models::Beta::Threads::TextDelta } - class TextDeltaBlock < OpenAI::BaseModel + class TextDeltaBlock < OpenAI::Internal::Type::BaseModel attr_accessor index: Integer attr_accessor type: :text diff --git a/sig/openai/models/chat/chat_completion.rbs b/sig/openai/models/chat/chat_completion.rbs index b0013010..54e0d554 100644 --- a/sig/openai/models/chat/chat_completion.rbs +++ b/sig/openai/models/chat/chat_completion.rbs @@ -15,7 +15,7 @@ module OpenAI usage: OpenAI::Models::CompletionUsage } - class ChatCompletion < OpenAI::BaseModel + class ChatCompletion < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor choices: ::Array[OpenAI::Models::Chat::ChatCompletion::Choice] @@ -59,7 +59,7 @@ module OpenAI message: OpenAI::Models::Chat::ChatCompletionMessage } - class Choice < OpenAI::BaseModel + class Choice < OpenAI::Internal::Type::BaseModel attr_accessor finish_reason: OpenAI::Models::Chat::ChatCompletion::Choice::finish_reason attr_accessor index: Integer @@ -81,7 +81,7 @@ module OpenAI :stop | :length | :tool_calls | :content_filter | :function_call module FinishReason - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum STOP: :stop LENGTH: :length @@ -98,7 +98,7 @@ module OpenAI refusal: ::Array[OpenAI::Models::Chat::ChatCompletionTokenLogprob]? } - class Logprobs < OpenAI::BaseModel + class Logprobs < OpenAI::Internal::Type::BaseModel attr_accessor content: ::Array[OpenAI::Models::Chat::ChatCompletionTokenLogprob]? attr_accessor refusal: ::Array[OpenAI::Models::Chat::ChatCompletionTokenLogprob]? @@ -115,7 +115,7 @@ module OpenAI type service_tier = :scale | :default module ServiceTier - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum SCALE: :scale DEFAULT: :default diff --git a/sig/openai/models/chat/chat_completion_assistant_message_param.rbs b/sig/openai/models/chat/chat_completion_assistant_message_param.rbs index 968d73be..4f92813c 100644 --- a/sig/openai/models/chat/chat_completion_assistant_message_param.rbs +++ b/sig/openai/models/chat/chat_completion_assistant_message_param.rbs @@ -14,7 +14,7 @@ module OpenAI tool_calls: ::Array[OpenAI::Models::Chat::ChatCompletionMessageToolCall] } - class ChatCompletionAssistantMessageParam < OpenAI::BaseModel + class ChatCompletionAssistantMessageParam < OpenAI::Internal::Type::BaseModel attr_accessor role: :assistant attr_accessor audio: OpenAI::Models::Chat::ChatCompletionAssistantMessageParam::Audio? @@ -49,7 +49,7 @@ module OpenAI type audio = { id: String } - class Audio < OpenAI::BaseModel + class Audio < OpenAI::Internal::Type::BaseModel attr_accessor id: String def initialize: (id: String) -> void @@ -62,14 +62,14 @@ module OpenAI | ::Array[OpenAI::Models::Chat::ChatCompletionAssistantMessageParam::Content::array_of_content_part] module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union type array_of_content_part = OpenAI::Models::Chat::ChatCompletionContentPartText | OpenAI::Models::Chat::ChatCompletionContentPartRefusal module ArrayOfContentPart - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Chat::ChatCompletionContentPartText, OpenAI::Models::Chat::ChatCompletionContentPartRefusal] end @@ -84,7 +84,7 @@ module OpenAI type function_call = { arguments: String, name: String } - class FunctionCall < OpenAI::BaseModel + class FunctionCall < OpenAI::Internal::Type::BaseModel attr_accessor arguments: String attr_accessor name: String diff --git a/sig/openai/models/chat/chat_completion_audio.rbs b/sig/openai/models/chat/chat_completion_audio.rbs index a3b3cda4..18d4ba70 100644 --- a/sig/openai/models/chat/chat_completion_audio.rbs +++ b/sig/openai/models/chat/chat_completion_audio.rbs @@ -6,7 +6,7 @@ module OpenAI type chat_completion_audio = { id: String, data: String, expires_at: Integer, transcript: String } - class ChatCompletionAudio < OpenAI::BaseModel + class ChatCompletionAudio < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor data: String diff --git a/sig/openai/models/chat/chat_completion_audio_param.rbs b/sig/openai/models/chat/chat_completion_audio_param.rbs index dc4be41d..3fb9c17d 100644 --- a/sig/openai/models/chat/chat_completion_audio_param.rbs +++ b/sig/openai/models/chat/chat_completion_audio_param.rbs @@ -9,7 +9,7 @@ module OpenAI voice: OpenAI::Models::Chat::ChatCompletionAudioParam::voice } - class ChatCompletionAudioParam < OpenAI::BaseModel + class ChatCompletionAudioParam < OpenAI::Internal::Type::BaseModel attr_accessor format_: OpenAI::Models::Chat::ChatCompletionAudioParam::format_ attr_accessor voice: OpenAI::Models::Chat::ChatCompletionAudioParam::voice @@ -24,7 +24,7 @@ module OpenAI type format_ = :wav | :mp3 | :flac | :opus | :pcm16 module Format - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum WAV: :wav MP3: :mp3 @@ -50,7 +50,7 @@ module OpenAI | :verse module Voice - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, (:alloy | :ash diff --git a/sig/openai/models/chat/chat_completion_chunk.rbs b/sig/openai/models/chat/chat_completion_chunk.rbs index 1ec7aadf..7cba70ad 100644 --- a/sig/openai/models/chat/chat_completion_chunk.rbs +++ b/sig/openai/models/chat/chat_completion_chunk.rbs @@ -15,7 +15,7 @@ module OpenAI usage: OpenAI::Models::CompletionUsage? } - class ChatCompletionChunk < OpenAI::BaseModel + class ChatCompletionChunk < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor choices: ::Array[OpenAI::Models::Chat::ChatCompletionChunk::Choice] @@ -55,7 +55,7 @@ module OpenAI logprobs: OpenAI::Models::Chat::ChatCompletionChunk::Choice::Logprobs? } - class Choice < OpenAI::BaseModel + class Choice < OpenAI::Internal::Type::BaseModel attr_accessor delta: OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta attr_accessor finish_reason: OpenAI::Models::Chat::ChatCompletionChunk::Choice::finish_reason? @@ -82,7 +82,7 @@ module OpenAI tool_calls: ::Array[OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta::ToolCall] } - class Delta < OpenAI::BaseModel + class Delta < OpenAI::Internal::Type::BaseModel attr_accessor content: String? attr_reader function_call: OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta::FunctionCall? @@ -117,7 +117,7 @@ module OpenAI type function_call = { arguments: String, name: String } - class FunctionCall < OpenAI::BaseModel + class FunctionCall < OpenAI::Internal::Type::BaseModel attr_reader arguments: String? def arguments=: (String) -> String @@ -134,7 +134,7 @@ module OpenAI type role = :developer | :system | :user | :assistant | :tool module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum DEVELOPER: :developer SYSTEM: :system @@ -153,7 +153,7 @@ module OpenAI type: OpenAI::Models::Chat::ChatCompletionChunk::Choice::Delta::ToolCall::type_ } - class ToolCall < OpenAI::BaseModel + class ToolCall < OpenAI::Internal::Type::BaseModel attr_accessor index: Integer attr_reader id: String? @@ -183,7 +183,7 @@ module OpenAI type function = { arguments: String, name: String } - class Function < OpenAI::BaseModel + class Function < OpenAI::Internal::Type::BaseModel attr_reader arguments: String? def arguments=: (String) -> String @@ -200,7 +200,7 @@ module OpenAI type type_ = :function module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum FUNCTION: :function @@ -213,7 +213,7 @@ module OpenAI :stop | :length | :tool_calls | :content_filter | :function_call module FinishReason - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum STOP: :stop LENGTH: :length @@ -230,7 +230,7 @@ module OpenAI refusal: ::Array[OpenAI::Models::Chat::ChatCompletionTokenLogprob]? } - class Logprobs < OpenAI::BaseModel + class Logprobs < OpenAI::Internal::Type::BaseModel attr_accessor content: ::Array[OpenAI::Models::Chat::ChatCompletionTokenLogprob]? attr_accessor refusal: ::Array[OpenAI::Models::Chat::ChatCompletionTokenLogprob]? @@ -247,7 +247,7 @@ module OpenAI type service_tier = :scale | :default module ServiceTier - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum SCALE: :scale DEFAULT: :default diff --git a/sig/openai/models/chat/chat_completion_content_part.rbs b/sig/openai/models/chat/chat_completion_content_part.rbs index df907b34..470cda35 100644 --- a/sig/openai/models/chat/chat_completion_content_part.rbs +++ b/sig/openai/models/chat/chat_completion_content_part.rbs @@ -10,7 +10,7 @@ module OpenAI | OpenAI::Models::Chat::ChatCompletionContentPart::File module ChatCompletionContentPart - extend OpenAI::Union + extend OpenAI::Internal::Type::Union type file = { @@ -18,7 +18,7 @@ module OpenAI type: :file } - class File < OpenAI::BaseModel + class File < OpenAI::Internal::Type::BaseModel attr_accessor file: OpenAI::Models::Chat::ChatCompletionContentPart::File::File attr_accessor type: :file @@ -32,7 +32,7 @@ module OpenAI type file = { file_data: String, file_id: String, filename: String } - class File < OpenAI::BaseModel + class File < OpenAI::Internal::Type::BaseModel attr_reader file_data: String? def file_data=: (String) -> String diff --git a/sig/openai/models/chat/chat_completion_content_part_image.rbs b/sig/openai/models/chat/chat_completion_content_part_image.rbs index d8ca22d9..6aa104fb 100644 --- a/sig/openai/models/chat/chat_completion_content_part_image.rbs +++ b/sig/openai/models/chat/chat_completion_content_part_image.rbs @@ -9,7 +9,7 @@ module OpenAI type: :image_url } - class ChatCompletionContentPartImage < OpenAI::BaseModel + class ChatCompletionContentPartImage < OpenAI::Internal::Type::BaseModel attr_accessor image_url: OpenAI::Models::Chat::ChatCompletionContentPartImage::ImageURL attr_accessor type: :image_url @@ -27,7 +27,7 @@ module OpenAI detail: OpenAI::Models::Chat::ChatCompletionContentPartImage::ImageURL::detail } - class ImageURL < OpenAI::BaseModel + class ImageURL < OpenAI::Internal::Type::BaseModel attr_accessor url: String attr_reader detail: OpenAI::Models::Chat::ChatCompletionContentPartImage::ImageURL::detail? @@ -46,7 +46,7 @@ module OpenAI type detail = :auto | :low | :high module Detail - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO: :auto LOW: :low diff --git a/sig/openai/models/chat/chat_completion_content_part_input_audio.rbs b/sig/openai/models/chat/chat_completion_content_part_input_audio.rbs index 8d40f203..e3267b32 100644 --- a/sig/openai/models/chat/chat_completion_content_part_input_audio.rbs +++ b/sig/openai/models/chat/chat_completion_content_part_input_audio.rbs @@ -9,7 +9,7 @@ module OpenAI type: :input_audio } - class ChatCompletionContentPartInputAudio < OpenAI::BaseModel + class ChatCompletionContentPartInputAudio < OpenAI::Internal::Type::BaseModel attr_accessor input_audio: OpenAI::Models::Chat::ChatCompletionContentPartInputAudio::InputAudio attr_accessor type: :input_audio @@ -27,7 +27,7 @@ module OpenAI format_: OpenAI::Models::Chat::ChatCompletionContentPartInputAudio::InputAudio::format_ } - class InputAudio < OpenAI::BaseModel + class InputAudio < OpenAI::Internal::Type::BaseModel attr_accessor data: String attr_accessor format_: OpenAI::Models::Chat::ChatCompletionContentPartInputAudio::InputAudio::format_ @@ -42,7 +42,7 @@ module OpenAI type format_ = :wav | :mp3 module Format - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum WAV: :wav MP3: :mp3 diff --git a/sig/openai/models/chat/chat_completion_content_part_refusal.rbs b/sig/openai/models/chat/chat_completion_content_part_refusal.rbs index e715e480..e7b62c6b 100644 --- a/sig/openai/models/chat/chat_completion_content_part_refusal.rbs +++ b/sig/openai/models/chat/chat_completion_content_part_refusal.rbs @@ -6,7 +6,7 @@ module OpenAI type chat_completion_content_part_refusal = { refusal: String, type: :refusal } - class ChatCompletionContentPartRefusal < OpenAI::BaseModel + class ChatCompletionContentPartRefusal < OpenAI::Internal::Type::BaseModel attr_accessor refusal: String attr_accessor type: :refusal diff --git a/sig/openai/models/chat/chat_completion_content_part_text.rbs b/sig/openai/models/chat/chat_completion_content_part_text.rbs index 638f5e4e..b8a43400 100644 --- a/sig/openai/models/chat/chat_completion_content_part_text.rbs +++ b/sig/openai/models/chat/chat_completion_content_part_text.rbs @@ -5,7 +5,7 @@ module OpenAI module Chat type chat_completion_content_part_text = { text: String, type: :text } - class ChatCompletionContentPartText < OpenAI::BaseModel + class ChatCompletionContentPartText < OpenAI::Internal::Type::BaseModel attr_accessor text: String attr_accessor type: :text diff --git a/sig/openai/models/chat/chat_completion_deleted.rbs b/sig/openai/models/chat/chat_completion_deleted.rbs index e776fc71..6d7e613c 100644 --- a/sig/openai/models/chat/chat_completion_deleted.rbs +++ b/sig/openai/models/chat/chat_completion_deleted.rbs @@ -6,7 +6,7 @@ module OpenAI type chat_completion_deleted = { id: String, deleted: bool, object: :"chat.completion.deleted" } - class ChatCompletionDeleted < OpenAI::BaseModel + class ChatCompletionDeleted < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor deleted: bool diff --git a/sig/openai/models/chat/chat_completion_developer_message_param.rbs b/sig/openai/models/chat/chat_completion_developer_message_param.rbs index 9bd2727c..b0dfdf86 100644 --- a/sig/openai/models/chat/chat_completion_developer_message_param.rbs +++ b/sig/openai/models/chat/chat_completion_developer_message_param.rbs @@ -10,7 +10,7 @@ module OpenAI name: String } - class ChatCompletionDeveloperMessageParam < OpenAI::BaseModel + class ChatCompletionDeveloperMessageParam < OpenAI::Internal::Type::BaseModel attr_accessor content: OpenAI::Models::Chat::ChatCompletionDeveloperMessageParam::content attr_accessor role: :developer @@ -31,7 +31,7 @@ module OpenAI String | ::Array[OpenAI::Models::Chat::ChatCompletionContentPartText] module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, ::Array[OpenAI::Models::Chat::ChatCompletionContentPartText]] diff --git a/sig/openai/models/chat/chat_completion_function_call_option.rbs b/sig/openai/models/chat/chat_completion_function_call_option.rbs index e7067a87..fd033d17 100644 --- a/sig/openai/models/chat/chat_completion_function_call_option.rbs +++ b/sig/openai/models/chat/chat_completion_function_call_option.rbs @@ -5,7 +5,7 @@ module OpenAI module Chat type chat_completion_function_call_option = { name: String } - class ChatCompletionFunctionCallOption < OpenAI::BaseModel + class ChatCompletionFunctionCallOption < OpenAI::Internal::Type::BaseModel attr_accessor name: String def initialize: (name: String) -> void diff --git a/sig/openai/models/chat/chat_completion_function_message_param.rbs b/sig/openai/models/chat/chat_completion_function_message_param.rbs index fad91c63..e83ffd82 100644 --- a/sig/openai/models/chat/chat_completion_function_message_param.rbs +++ b/sig/openai/models/chat/chat_completion_function_message_param.rbs @@ -6,7 +6,7 @@ module OpenAI type chat_completion_function_message_param = { content: String?, name: String, role: :function } - class ChatCompletionFunctionMessageParam < OpenAI::BaseModel + class ChatCompletionFunctionMessageParam < OpenAI::Internal::Type::BaseModel attr_accessor content: String? attr_accessor name: String diff --git a/sig/openai/models/chat/chat_completion_message.rbs b/sig/openai/models/chat/chat_completion_message.rbs index d3668749..7fa0b355 100644 --- a/sig/openai/models/chat/chat_completion_message.rbs +++ b/sig/openai/models/chat/chat_completion_message.rbs @@ -14,7 +14,7 @@ module OpenAI tool_calls: ::Array[OpenAI::Models::Chat::ChatCompletionMessageToolCall] } - class ChatCompletionMessage < OpenAI::BaseModel + class ChatCompletionMessage < OpenAI::Internal::Type::BaseModel attr_accessor content: String? attr_accessor refusal: String? @@ -59,7 +59,7 @@ module OpenAI url_citation: OpenAI::Models::Chat::ChatCompletionMessage::Annotation::URLCitation } - class Annotation < OpenAI::BaseModel + class Annotation < OpenAI::Internal::Type::BaseModel attr_accessor type: :url_citation attr_accessor url_citation: OpenAI::Models::Chat::ChatCompletionMessage::Annotation::URLCitation @@ -79,7 +79,7 @@ module OpenAI url: String } - class URLCitation < OpenAI::BaseModel + class URLCitation < OpenAI::Internal::Type::BaseModel attr_accessor end_index: Integer attr_accessor start_index: Integer @@ -101,7 +101,7 @@ module OpenAI type function_call = { arguments: String, name: String } - class FunctionCall < OpenAI::BaseModel + class FunctionCall < OpenAI::Internal::Type::BaseModel attr_accessor arguments: String attr_accessor name: String diff --git a/sig/openai/models/chat/chat_completion_message_param.rbs b/sig/openai/models/chat/chat_completion_message_param.rbs index 557d3c2c..e30fd658 100644 --- a/sig/openai/models/chat/chat_completion_message_param.rbs +++ b/sig/openai/models/chat/chat_completion_message_param.rbs @@ -12,7 +12,7 @@ module OpenAI | OpenAI::Models::Chat::ChatCompletionFunctionMessageParam module ChatCompletionMessageParam - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Chat::ChatCompletionDeveloperMessageParam, OpenAI::Models::Chat::ChatCompletionSystemMessageParam, OpenAI::Models::Chat::ChatCompletionUserMessageParam, OpenAI::Models::Chat::ChatCompletionAssistantMessageParam, OpenAI::Models::Chat::ChatCompletionToolMessageParam, OpenAI::Models::Chat::ChatCompletionFunctionMessageParam] end diff --git a/sig/openai/models/chat/chat_completion_message_tool_call.rbs b/sig/openai/models/chat/chat_completion_message_tool_call.rbs index cd147e2d..ca90663e 100644 --- a/sig/openai/models/chat/chat_completion_message_tool_call.rbs +++ b/sig/openai/models/chat/chat_completion_message_tool_call.rbs @@ -10,7 +10,7 @@ module OpenAI type: :function } - class ChatCompletionMessageToolCall < OpenAI::BaseModel + class ChatCompletionMessageToolCall < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor function: OpenAI::Models::Chat::ChatCompletionMessageToolCall::Function @@ -27,7 +27,7 @@ module OpenAI type function = { arguments: String, name: String } - class Function < OpenAI::BaseModel + class Function < OpenAI::Internal::Type::BaseModel attr_accessor arguments: String attr_accessor name: String diff --git a/sig/openai/models/chat/chat_completion_modality.rbs b/sig/openai/models/chat/chat_completion_modality.rbs index b634c024..c0b279d7 100644 --- a/sig/openai/models/chat/chat_completion_modality.rbs +++ b/sig/openai/models/chat/chat_completion_modality.rbs @@ -6,7 +6,7 @@ module OpenAI type chat_completion_modality = :text | :audio module ChatCompletionModality - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT: :text AUDIO: :audio diff --git a/sig/openai/models/chat/chat_completion_named_tool_choice.rbs b/sig/openai/models/chat/chat_completion_named_tool_choice.rbs index 62feb1bf..77b79236 100644 --- a/sig/openai/models/chat/chat_completion_named_tool_choice.rbs +++ b/sig/openai/models/chat/chat_completion_named_tool_choice.rbs @@ -9,7 +9,7 @@ module OpenAI type: :function } - class ChatCompletionNamedToolChoice < OpenAI::BaseModel + class ChatCompletionNamedToolChoice < OpenAI::Internal::Type::BaseModel attr_accessor function: OpenAI::Models::Chat::ChatCompletionNamedToolChoice::Function attr_accessor type: :function @@ -23,7 +23,7 @@ module OpenAI type function = { name: String } - class Function < OpenAI::BaseModel + class Function < OpenAI::Internal::Type::BaseModel attr_accessor name: String def initialize: (name: String) -> void diff --git a/sig/openai/models/chat/chat_completion_prediction_content.rbs b/sig/openai/models/chat/chat_completion_prediction_content.rbs index cbf6347c..1f5c681a 100644 --- a/sig/openai/models/chat/chat_completion_prediction_content.rbs +++ b/sig/openai/models/chat/chat_completion_prediction_content.rbs @@ -9,7 +9,7 @@ module OpenAI type: :content } - class ChatCompletionPredictionContent < OpenAI::BaseModel + class ChatCompletionPredictionContent < OpenAI::Internal::Type::BaseModel attr_accessor content: OpenAI::Models::Chat::ChatCompletionPredictionContent::content attr_accessor type: :content @@ -25,7 +25,7 @@ module OpenAI String | ::Array[OpenAI::Models::Chat::ChatCompletionContentPartText] module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, ::Array[OpenAI::Models::Chat::ChatCompletionContentPartText]] diff --git a/sig/openai/models/chat/chat_completion_role.rbs b/sig/openai/models/chat/chat_completion_role.rbs index 4744870b..96104afd 100644 --- a/sig/openai/models/chat/chat_completion_role.rbs +++ b/sig/openai/models/chat/chat_completion_role.rbs @@ -7,7 +7,7 @@ module OpenAI :developer | :system | :user | :assistant | :tool | :function module ChatCompletionRole - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum DEVELOPER: :developer SYSTEM: :system diff --git a/sig/openai/models/chat/chat_completion_stream_options.rbs b/sig/openai/models/chat/chat_completion_stream_options.rbs index 7fed3536..bf86a211 100644 --- a/sig/openai/models/chat/chat_completion_stream_options.rbs +++ b/sig/openai/models/chat/chat_completion_stream_options.rbs @@ -5,7 +5,7 @@ module OpenAI module Chat type chat_completion_stream_options = { include_usage: bool } - class ChatCompletionStreamOptions < OpenAI::BaseModel + class ChatCompletionStreamOptions < OpenAI::Internal::Type::BaseModel attr_reader include_usage: bool? def include_usage=: (bool) -> bool diff --git a/sig/openai/models/chat/chat_completion_system_message_param.rbs b/sig/openai/models/chat/chat_completion_system_message_param.rbs index 56a477ec..12811868 100644 --- a/sig/openai/models/chat/chat_completion_system_message_param.rbs +++ b/sig/openai/models/chat/chat_completion_system_message_param.rbs @@ -10,7 +10,7 @@ module OpenAI name: String } - class ChatCompletionSystemMessageParam < OpenAI::BaseModel + class ChatCompletionSystemMessageParam < OpenAI::Internal::Type::BaseModel attr_accessor content: OpenAI::Models::Chat::ChatCompletionSystemMessageParam::content attr_accessor role: :system @@ -31,7 +31,7 @@ module OpenAI String | ::Array[OpenAI::Models::Chat::ChatCompletionContentPartText] module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, ::Array[OpenAI::Models::Chat::ChatCompletionContentPartText]] diff --git a/sig/openai/models/chat/chat_completion_token_logprob.rbs b/sig/openai/models/chat/chat_completion_token_logprob.rbs index a076afef..968dd8a5 100644 --- a/sig/openai/models/chat/chat_completion_token_logprob.rbs +++ b/sig/openai/models/chat/chat_completion_token_logprob.rbs @@ -11,7 +11,7 @@ module OpenAI top_logprobs: ::Array[OpenAI::Models::Chat::ChatCompletionTokenLogprob::TopLogprob] } - class ChatCompletionTokenLogprob < OpenAI::BaseModel + class ChatCompletionTokenLogprob < OpenAI::Internal::Type::BaseModel attr_accessor token: String attr_accessor bytes: ::Array[Integer]? @@ -32,7 +32,7 @@ module OpenAI type top_logprob = { token: String, bytes: ::Array[Integer]?, logprob: Float } - class TopLogprob < OpenAI::BaseModel + class TopLogprob < OpenAI::Internal::Type::BaseModel attr_accessor token: String attr_accessor bytes: ::Array[Integer]? diff --git a/sig/openai/models/chat/chat_completion_tool.rbs b/sig/openai/models/chat/chat_completion_tool.rbs index d465043d..1fd217b1 100644 --- a/sig/openai/models/chat/chat_completion_tool.rbs +++ b/sig/openai/models/chat/chat_completion_tool.rbs @@ -6,7 +6,7 @@ module OpenAI type chat_completion_tool = { function: OpenAI::Models::FunctionDefinition, type: :function } - class ChatCompletionTool < OpenAI::BaseModel + class ChatCompletionTool < OpenAI::Internal::Type::BaseModel attr_accessor function: OpenAI::Models::FunctionDefinition attr_accessor type: :function diff --git a/sig/openai/models/chat/chat_completion_tool_choice_option.rbs b/sig/openai/models/chat/chat_completion_tool_choice_option.rbs index 114b7588..aadc1e00 100644 --- a/sig/openai/models/chat/chat_completion_tool_choice_option.rbs +++ b/sig/openai/models/chat/chat_completion_tool_choice_option.rbs @@ -8,12 +8,12 @@ module OpenAI | OpenAI::Models::Chat::ChatCompletionNamedToolChoice module ChatCompletionToolChoiceOption - extend OpenAI::Union + extend OpenAI::Internal::Type::Union type auto = :none | :auto | :required module Auto - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum NONE: :none AUTO: :auto diff --git a/sig/openai/models/chat/chat_completion_tool_message_param.rbs b/sig/openai/models/chat/chat_completion_tool_message_param.rbs index e0e31f90..7808c0a9 100644 --- a/sig/openai/models/chat/chat_completion_tool_message_param.rbs +++ b/sig/openai/models/chat/chat_completion_tool_message_param.rbs @@ -10,7 +10,7 @@ module OpenAI tool_call_id: String } - class ChatCompletionToolMessageParam < OpenAI::BaseModel + class ChatCompletionToolMessageParam < OpenAI::Internal::Type::BaseModel attr_accessor content: OpenAI::Models::Chat::ChatCompletionToolMessageParam::content attr_accessor role: :tool @@ -29,7 +29,7 @@ module OpenAI String | ::Array[OpenAI::Models::Chat::ChatCompletionContentPartText] module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, ::Array[OpenAI::Models::Chat::ChatCompletionContentPartText]] diff --git a/sig/openai/models/chat/chat_completion_user_message_param.rbs b/sig/openai/models/chat/chat_completion_user_message_param.rbs index 308744d9..ea9b16d6 100644 --- a/sig/openai/models/chat/chat_completion_user_message_param.rbs +++ b/sig/openai/models/chat/chat_completion_user_message_param.rbs @@ -10,7 +10,7 @@ module OpenAI name: String } - class ChatCompletionUserMessageParam < OpenAI::BaseModel + class ChatCompletionUserMessageParam < OpenAI::Internal::Type::BaseModel attr_accessor content: OpenAI::Models::Chat::ChatCompletionUserMessageParam::content attr_accessor role: :user @@ -31,7 +31,7 @@ module OpenAI String | ::Array[OpenAI::Models::Chat::chat_completion_content_part] module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, ::Array[OpenAI::Models::Chat::chat_completion_content_part]] diff --git a/sig/openai/models/chat/completion_create_params.rbs b/sig/openai/models/chat/completion_create_params.rbs index 1515d670..d10cf21d 100644 --- a/sig/openai/models/chat/completion_create_params.rbs +++ b/sig/openai/models/chat/completion_create_params.rbs @@ -36,7 +36,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class CompletionCreateParams < OpenAI::BaseModel + class CompletionCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -167,7 +167,7 @@ module OpenAI type model = String | OpenAI::Models::chat_model module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, OpenAI::Models::chat_model] end @@ -177,12 +177,12 @@ module OpenAI | OpenAI::Models::Chat::ChatCompletionFunctionCallOption module FunctionCall - extend OpenAI::Union + extend OpenAI::Internal::Type::Union type function_call_mode = :none | :auto module FunctionCallMode - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum NONE: :none AUTO: :auto @@ -200,7 +200,7 @@ module OpenAI parameters: OpenAI::Models::function_parameters } - class Function < OpenAI::BaseModel + class Function < OpenAI::Internal::Type::BaseModel attr_accessor name: String attr_reader description: String? @@ -225,7 +225,7 @@ module OpenAI type modality = :text | :audio module Modality - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT: :text AUDIO: :audio @@ -239,7 +239,7 @@ module OpenAI | OpenAI::Models::ResponseFormatJSONObject module ResponseFormat - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::ResponseFormatText, OpenAI::Models::ResponseFormatJSONSchema, OpenAI::Models::ResponseFormatJSONObject] end @@ -247,7 +247,7 @@ module OpenAI type service_tier = :auto | :default module ServiceTier - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO: :auto DEFAULT: :default @@ -258,7 +258,7 @@ module OpenAI type stop = (String | ::Array[String])? module Stop - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, ::Array[String]] @@ -273,7 +273,7 @@ module OpenAI user_location: OpenAI::Models::Chat::CompletionCreateParams::WebSearchOptions::UserLocation? } - class WebSearchOptions < OpenAI::BaseModel + class WebSearchOptions < OpenAI::Internal::Type::BaseModel attr_reader search_context_size: OpenAI::Models::Chat::CompletionCreateParams::WebSearchOptions::search_context_size? def search_context_size=: ( @@ -292,7 +292,7 @@ module OpenAI type search_context_size = :low | :medium | :high module SearchContextSize - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum LOW: :low MEDIUM: :medium @@ -307,7 +307,7 @@ module OpenAI type: :approximate } - class UserLocation < OpenAI::BaseModel + class UserLocation < OpenAI::Internal::Type::BaseModel attr_accessor approximate: OpenAI::Models::Chat::CompletionCreateParams::WebSearchOptions::UserLocation::Approximate attr_accessor type: :approximate @@ -327,7 +327,7 @@ module OpenAI timezone: String } - class Approximate < OpenAI::BaseModel + class Approximate < OpenAI::Internal::Type::BaseModel attr_reader city: String? def city=: (String) -> String diff --git a/sig/openai/models/chat/completion_delete_params.rbs b/sig/openai/models/chat/completion_delete_params.rbs index e4206b21..324fe253 100644 --- a/sig/openai/models/chat/completion_delete_params.rbs +++ b/sig/openai/models/chat/completion_delete_params.rbs @@ -4,7 +4,7 @@ module OpenAI type completion_delete_params = { } & OpenAI::Internal::Type::request_parameters - class CompletionDeleteParams < OpenAI::BaseModel + class CompletionDeleteParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/chat/completion_list_params.rbs b/sig/openai/models/chat/completion_list_params.rbs index 7f171502..e7bbb7d5 100644 --- a/sig/openai/models/chat/completion_list_params.rbs +++ b/sig/openai/models/chat/completion_list_params.rbs @@ -11,7 +11,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class CompletionListParams < OpenAI::BaseModel + class CompletionListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -49,7 +49,7 @@ module OpenAI type order = :asc | :desc module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASC: :asc DESC: :desc diff --git a/sig/openai/models/chat/completion_retrieve_params.rbs b/sig/openai/models/chat/completion_retrieve_params.rbs index 3c64e3fb..7d97ef11 100644 --- a/sig/openai/models/chat/completion_retrieve_params.rbs +++ b/sig/openai/models/chat/completion_retrieve_params.rbs @@ -4,7 +4,7 @@ module OpenAI type completion_retrieve_params = { } & OpenAI::Internal::Type::request_parameters - class CompletionRetrieveParams < OpenAI::BaseModel + class CompletionRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/chat/completion_update_params.rbs b/sig/openai/models/chat/completion_update_params.rbs index dbfdd87d..dc63a22a 100644 --- a/sig/openai/models/chat/completion_update_params.rbs +++ b/sig/openai/models/chat/completion_update_params.rbs @@ -5,7 +5,7 @@ module OpenAI { metadata: OpenAI::Models::metadata? } & OpenAI::Internal::Type::request_parameters - class CompletionUpdateParams < OpenAI::BaseModel + class CompletionUpdateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/chat/completions/message_list_params.rbs b/sig/openai/models/chat/completions/message_list_params.rbs index 1d5f30a2..73db541f 100644 --- a/sig/openai/models/chat/completions/message_list_params.rbs +++ b/sig/openai/models/chat/completions/message_list_params.rbs @@ -10,7 +10,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class MessageListParams < OpenAI::BaseModel + class MessageListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -40,7 +40,7 @@ module OpenAI type order = :asc | :desc module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASC: :asc DESC: :desc diff --git a/sig/openai/models/chat_model.rbs b/sig/openai/models/chat_model.rbs index d0aa3240..0f571b30 100644 --- a/sig/openai/models/chat_model.rbs +++ b/sig/openai/models/chat_model.rbs @@ -46,7 +46,7 @@ module OpenAI | :"gpt-3.5-turbo-16k-0613" module ChatModel - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum O3_MINI: :"o3-mini" O3_MINI_2025_01_31: :"o3-mini-2025-01-31" diff --git a/sig/openai/models/comparison_filter.rbs b/sig/openai/models/comparison_filter.rbs index 8d14f63b..e4a99bcf 100644 --- a/sig/openai/models/comparison_filter.rbs +++ b/sig/openai/models/comparison_filter.rbs @@ -7,7 +7,7 @@ module OpenAI value: OpenAI::Models::ComparisonFilter::value } - class ComparisonFilter < OpenAI::BaseModel + class ComparisonFilter < OpenAI::Internal::Type::BaseModel attr_accessor key: String attr_accessor type: OpenAI::Models::ComparisonFilter::type_ @@ -25,7 +25,7 @@ module OpenAI type type_ = :eq | :ne | :gt | :gte | :lt | :lte module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum EQ: :eq NE: :ne @@ -40,7 +40,7 @@ module OpenAI type value = String | Float | bool module Value - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, Float, bool] end diff --git a/sig/openai/models/completion.rbs b/sig/openai/models/completion.rbs index 0e189869..ed2ca5d6 100644 --- a/sig/openai/models/completion.rbs +++ b/sig/openai/models/completion.rbs @@ -11,7 +11,7 @@ module OpenAI usage: OpenAI::Models::CompletionUsage } - class Completion < OpenAI::BaseModel + class Completion < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor choices: ::Array[OpenAI::Models::CompletionChoice] diff --git a/sig/openai/models/completion_choice.rbs b/sig/openai/models/completion_choice.rbs index e506a455..5e50fc5f 100644 --- a/sig/openai/models/completion_choice.rbs +++ b/sig/openai/models/completion_choice.rbs @@ -8,7 +8,7 @@ module OpenAI text: String } - class CompletionChoice < OpenAI::BaseModel + class CompletionChoice < OpenAI::Internal::Type::BaseModel attr_accessor finish_reason: OpenAI::Models::CompletionChoice::finish_reason attr_accessor index: Integer @@ -29,7 +29,7 @@ module OpenAI type finish_reason = :stop | :length | :content_filter module FinishReason - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum STOP: :stop LENGTH: :length @@ -46,7 +46,7 @@ module OpenAI top_logprobs: ::Array[::Hash[Symbol, Float]] } - class Logprobs < OpenAI::BaseModel + class Logprobs < OpenAI::Internal::Type::BaseModel attr_reader text_offset: ::Array[Integer]? def text_offset=: (::Array[Integer]) -> ::Array[Integer] diff --git a/sig/openai/models/completion_create_params.rbs b/sig/openai/models/completion_create_params.rbs index 3bbe8da3..d4b9afbe 100644 --- a/sig/openai/models/completion_create_params.rbs +++ b/sig/openai/models/completion_create_params.rbs @@ -22,7 +22,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class CompletionCreateParams < OpenAI::BaseModel + class CompletionCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -89,7 +89,7 @@ module OpenAI String | :"gpt-3.5-turbo-instruct" | :"davinci-002" | :"babbage-002" module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, (:"gpt-3.5-turbo-instruct" | :"davinci-002" @@ -104,7 +104,7 @@ module OpenAI String | ::Array[String] | ::Array[Integer] | ::Array[::Array[Integer]] module Prompt - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, ::Array[String], ::Array[Integer], ::Array[::Array[Integer]]] @@ -124,7 +124,7 @@ module OpenAI type stop = (String | ::Array[String])? module Stop - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, ::Array[String]] diff --git a/sig/openai/models/completion_usage.rbs b/sig/openai/models/completion_usage.rbs index fe20a24b..5a98db8b 100644 --- a/sig/openai/models/completion_usage.rbs +++ b/sig/openai/models/completion_usage.rbs @@ -9,7 +9,7 @@ module OpenAI prompt_tokens_details: OpenAI::Models::CompletionUsage::PromptTokensDetails } - class CompletionUsage < OpenAI::BaseModel + class CompletionUsage < OpenAI::Internal::Type::BaseModel attr_accessor completion_tokens: Integer attr_accessor prompt_tokens: Integer @@ -46,7 +46,7 @@ module OpenAI rejected_prediction_tokens: Integer } - class CompletionTokensDetails < OpenAI::BaseModel + class CompletionTokensDetails < OpenAI::Internal::Type::BaseModel attr_reader accepted_prediction_tokens: Integer? def accepted_prediction_tokens=: (Integer) -> Integer @@ -76,7 +76,7 @@ module OpenAI type prompt_tokens_details = { audio_tokens: Integer, cached_tokens: Integer } - class PromptTokensDetails < OpenAI::BaseModel + class PromptTokensDetails < OpenAI::Internal::Type::BaseModel attr_reader audio_tokens: Integer? def audio_tokens=: (Integer) -> Integer diff --git a/sig/openai/models/compound_filter.rbs b/sig/openai/models/compound_filter.rbs index 211de484..2f98c0d3 100644 --- a/sig/openai/models/compound_filter.rbs +++ b/sig/openai/models/compound_filter.rbs @@ -6,7 +6,7 @@ module OpenAI type: OpenAI::Models::CompoundFilter::type_ } - class CompoundFilter < OpenAI::BaseModel + class CompoundFilter < OpenAI::Internal::Type::BaseModel attr_accessor filters: ::Array[OpenAI::Models::CompoundFilter::filter] attr_accessor type: OpenAI::Models::CompoundFilter::type_ @@ -21,7 +21,7 @@ module OpenAI type filter = OpenAI::Models::ComparisonFilter | top module Filter - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::ComparisonFilter, top] end @@ -29,7 +29,7 @@ module OpenAI type type_ = :and | :or module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AND: :and OR: :or diff --git a/sig/openai/models/create_embedding_response.rbs b/sig/openai/models/create_embedding_response.rbs index e9ae13dc..d9d813c3 100644 --- a/sig/openai/models/create_embedding_response.rbs +++ b/sig/openai/models/create_embedding_response.rbs @@ -8,7 +8,7 @@ module OpenAI usage: OpenAI::Models::CreateEmbeddingResponse::Usage } - class CreateEmbeddingResponse < OpenAI::BaseModel + class CreateEmbeddingResponse < OpenAI::Internal::Type::BaseModel attr_accessor data: ::Array[OpenAI::Models::Embedding] attr_accessor model: String @@ -28,7 +28,7 @@ module OpenAI type usage = { prompt_tokens: Integer, total_tokens: Integer } - class Usage < OpenAI::BaseModel + class Usage < OpenAI::Internal::Type::BaseModel attr_accessor prompt_tokens: Integer attr_accessor total_tokens: Integer diff --git a/sig/openai/models/embedding.rbs b/sig/openai/models/embedding.rbs index 08eb26a7..c0787d7b 100644 --- a/sig/openai/models/embedding.rbs +++ b/sig/openai/models/embedding.rbs @@ -3,7 +3,7 @@ module OpenAI type embedding = { embedding: ::Array[Float], index: Integer, object: :embedding } - class Embedding < OpenAI::BaseModel + class Embedding < OpenAI::Internal::Type::BaseModel attr_accessor embedding: ::Array[Float] attr_accessor index: Integer diff --git a/sig/openai/models/embedding_create_params.rbs b/sig/openai/models/embedding_create_params.rbs index 1c00db8b..fc247341 100644 --- a/sig/openai/models/embedding_create_params.rbs +++ b/sig/openai/models/embedding_create_params.rbs @@ -10,7 +10,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class EmbeddingCreateParams < OpenAI::BaseModel + class EmbeddingCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -47,7 +47,7 @@ module OpenAI String | ::Array[String] | ::Array[Integer] | ::Array[::Array[Integer]] module Input - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, ::Array[String], ::Array[Integer], ::Array[::Array[Integer]]] @@ -67,7 +67,7 @@ module OpenAI type model = String | OpenAI::Models::embedding_model module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, OpenAI::Models::embedding_model] end @@ -75,7 +75,7 @@ module OpenAI type encoding_format = :float | :base64 module EncodingFormat - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum FLOAT: :float BASE64: :base64 diff --git a/sig/openai/models/embedding_model.rbs b/sig/openai/models/embedding_model.rbs index 30fd2a8a..eebc0c7b 100644 --- a/sig/openai/models/embedding_model.rbs +++ b/sig/openai/models/embedding_model.rbs @@ -6,7 +6,7 @@ module OpenAI | :"text-embedding-3-large" module EmbeddingModel - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT_EMBEDDING_ADA_002: :"text-embedding-ada-002" TEXT_EMBEDDING_3_SMALL: :"text-embedding-3-small" diff --git a/sig/openai/models/error_object.rbs b/sig/openai/models/error_object.rbs index bd0ae3a7..0a526b8a 100644 --- a/sig/openai/models/error_object.rbs +++ b/sig/openai/models/error_object.rbs @@ -3,7 +3,7 @@ module OpenAI type error_object = { code: String?, message: String, param: String?, type: String } - class ErrorObject < OpenAI::BaseModel + class ErrorObject < OpenAI::Internal::Type::BaseModel attr_accessor code: String? attr_accessor message: String diff --git a/sig/openai/models/file_chunking_strategy.rbs b/sig/openai/models/file_chunking_strategy.rbs index 7ac3728a..ebe4dcb5 100644 --- a/sig/openai/models/file_chunking_strategy.rbs +++ b/sig/openai/models/file_chunking_strategy.rbs @@ -5,7 +5,7 @@ module OpenAI | OpenAI::Models::OtherFileChunkingStrategyObject module FileChunkingStrategy - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::StaticFileChunkingStrategyObject, OpenAI::Models::OtherFileChunkingStrategyObject] end diff --git a/sig/openai/models/file_chunking_strategy_param.rbs b/sig/openai/models/file_chunking_strategy_param.rbs index df5528b1..22233a43 100644 --- a/sig/openai/models/file_chunking_strategy_param.rbs +++ b/sig/openai/models/file_chunking_strategy_param.rbs @@ -5,7 +5,7 @@ module OpenAI | OpenAI::Models::StaticFileChunkingStrategyObjectParam module FileChunkingStrategyParam - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::AutoFileChunkingStrategyParam, OpenAI::Models::StaticFileChunkingStrategyObjectParam] end diff --git a/sig/openai/models/file_content_params.rbs b/sig/openai/models/file_content_params.rbs index c9adfe43..d0eb084a 100644 --- a/sig/openai/models/file_content_params.rbs +++ b/sig/openai/models/file_content_params.rbs @@ -2,7 +2,7 @@ module OpenAI module Models type file_content_params = { } & OpenAI::Internal::Type::request_parameters - class FileContentParams < OpenAI::BaseModel + class FileContentParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/file_create_params.rbs b/sig/openai/models/file_create_params.rbs index f08a1328..46d8450e 100644 --- a/sig/openai/models/file_create_params.rbs +++ b/sig/openai/models/file_create_params.rbs @@ -4,7 +4,7 @@ module OpenAI { file: (IO | StringIO), purpose: OpenAI::Models::file_purpose } & OpenAI::Internal::Type::request_parameters - class FileCreateParams < OpenAI::BaseModel + class FileCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/file_delete_params.rbs b/sig/openai/models/file_delete_params.rbs index 993174ef..bd9575f3 100644 --- a/sig/openai/models/file_delete_params.rbs +++ b/sig/openai/models/file_delete_params.rbs @@ -2,7 +2,7 @@ module OpenAI module Models type file_delete_params = { } & OpenAI::Internal::Type::request_parameters - class FileDeleteParams < OpenAI::BaseModel + class FileDeleteParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/file_deleted.rbs b/sig/openai/models/file_deleted.rbs index 04eef616..e4134f28 100644 --- a/sig/openai/models/file_deleted.rbs +++ b/sig/openai/models/file_deleted.rbs @@ -2,7 +2,7 @@ module OpenAI module Models type file_deleted = { id: String, deleted: bool, object: :file } - class FileDeleted < OpenAI::BaseModel + class FileDeleted < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor deleted: bool diff --git a/sig/openai/models/file_list_params.rbs b/sig/openai/models/file_list_params.rbs index 89b0392a..8cc477d7 100644 --- a/sig/openai/models/file_list_params.rbs +++ b/sig/openai/models/file_list_params.rbs @@ -9,7 +9,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class FileListParams < OpenAI::BaseModel + class FileListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -44,7 +44,7 @@ module OpenAI type order = :asc | :desc module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASC: :asc DESC: :desc diff --git a/sig/openai/models/file_object.rbs b/sig/openai/models/file_object.rbs index c4b2cd6b..2ea4ec93 100644 --- a/sig/openai/models/file_object.rbs +++ b/sig/openai/models/file_object.rbs @@ -13,7 +13,7 @@ module OpenAI status_details: String } - class FileObject < OpenAI::BaseModel + class FileObject < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor bytes: Integer @@ -60,7 +60,7 @@ module OpenAI | :vision module Purpose - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASSISTANTS: :assistants ASSISTANTS_OUTPUT: :assistants_output @@ -76,7 +76,7 @@ module OpenAI type status = :uploaded | :processed | :error module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum UPLOADED: :uploaded PROCESSED: :processed diff --git a/sig/openai/models/file_purpose.rbs b/sig/openai/models/file_purpose.rbs index 84067b2c..9b17621d 100644 --- a/sig/openai/models/file_purpose.rbs +++ b/sig/openai/models/file_purpose.rbs @@ -4,7 +4,7 @@ module OpenAI :assistants | :batch | :"fine-tune" | :vision | :user_data | :evals module FilePurpose - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASSISTANTS: :assistants BATCH: :batch diff --git a/sig/openai/models/file_retrieve_params.rbs b/sig/openai/models/file_retrieve_params.rbs index e81b21d8..818e9432 100644 --- a/sig/openai/models/file_retrieve_params.rbs +++ b/sig/openai/models/file_retrieve_params.rbs @@ -3,7 +3,7 @@ module OpenAI type file_retrieve_params = { } & OpenAI::Internal::Type::request_parameters - class FileRetrieveParams < OpenAI::BaseModel + class FileRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/fine_tuning/fine_tuning_job.rbs b/sig/openai/models/fine_tuning/fine_tuning_job.rbs index b939849a..8a66e5a9 100644 --- a/sig/openai/models/fine_tuning/fine_tuning_job.rbs +++ b/sig/openai/models/fine_tuning/fine_tuning_job.rbs @@ -26,7 +26,7 @@ module OpenAI method_: OpenAI::Models::FineTuning::FineTuningJob::Method } - class FineTuningJob < OpenAI::BaseModel + class FineTuningJob < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor created_at: Integer @@ -95,7 +95,7 @@ module OpenAI type error = { code: String, message: String, param: String? } - class Error < OpenAI::BaseModel + class Error < OpenAI::Internal::Type::BaseModel attr_accessor code: String attr_accessor message: String @@ -118,7 +118,7 @@ module OpenAI n_epochs: OpenAI::Models::FineTuning::FineTuningJob::Hyperparameters::n_epochs } - class Hyperparameters < OpenAI::BaseModel + class Hyperparameters < OpenAI::Internal::Type::BaseModel attr_reader batch_size: OpenAI::Models::FineTuning::FineTuningJob::Hyperparameters::batch_size? def batch_size=: ( @@ -148,7 +148,7 @@ module OpenAI type batch_size = :auto | Integer module BatchSize - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [:auto, Integer] end @@ -156,7 +156,7 @@ module OpenAI type learning_rate_multiplier = :auto | Float module LearningRateMultiplier - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [:auto, Float] end @@ -164,7 +164,7 @@ module OpenAI type n_epochs = :auto | Integer module NEpochs - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [:auto, Integer] end @@ -179,7 +179,7 @@ module OpenAI | :cancelled module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum VALIDATING_FILES: :validating_files QUEUED: :queued @@ -198,7 +198,7 @@ module OpenAI type: OpenAI::Models::FineTuning::FineTuningJob::Method::type_ } - class Method < OpenAI::BaseModel + class Method < OpenAI::Internal::Type::BaseModel attr_reader dpo: OpenAI::Models::FineTuning::FineTuningJob::Method::Dpo? def dpo=: ( @@ -230,7 +230,7 @@ module OpenAI hyperparameters: OpenAI::Models::FineTuning::FineTuningJob::Method::Dpo::Hyperparameters } - class Dpo < OpenAI::BaseModel + class Dpo < OpenAI::Internal::Type::BaseModel attr_reader hyperparameters: OpenAI::Models::FineTuning::FineTuningJob::Method::Dpo::Hyperparameters? def hyperparameters=: ( @@ -251,7 +251,7 @@ module OpenAI n_epochs: OpenAI::Models::FineTuning::FineTuningJob::Method::Dpo::Hyperparameters::n_epochs } - class Hyperparameters < OpenAI::BaseModel + class Hyperparameters < OpenAI::Internal::Type::BaseModel attr_reader batch_size: OpenAI::Models::FineTuning::FineTuningJob::Method::Dpo::Hyperparameters::batch_size? def batch_size=: ( @@ -288,7 +288,7 @@ module OpenAI type batch_size = :auto | Integer module BatchSize - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [:auto, Integer] end @@ -296,7 +296,7 @@ module OpenAI type beta = :auto | Float module Beta - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [:auto, Float] end @@ -304,7 +304,7 @@ module OpenAI type learning_rate_multiplier = :auto | Float module LearningRateMultiplier - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [:auto, Float] end @@ -312,7 +312,7 @@ module OpenAI type n_epochs = :auto | Integer module NEpochs - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [:auto, Integer] end @@ -324,7 +324,7 @@ module OpenAI hyperparameters: OpenAI::Models::FineTuning::FineTuningJob::Method::Supervised::Hyperparameters } - class Supervised < OpenAI::BaseModel + class Supervised < OpenAI::Internal::Type::BaseModel attr_reader hyperparameters: OpenAI::Models::FineTuning::FineTuningJob::Method::Supervised::Hyperparameters? def hyperparameters=: ( @@ -344,7 +344,7 @@ module OpenAI n_epochs: OpenAI::Models::FineTuning::FineTuningJob::Method::Supervised::Hyperparameters::n_epochs } - class Hyperparameters < OpenAI::BaseModel + class Hyperparameters < OpenAI::Internal::Type::BaseModel attr_reader batch_size: OpenAI::Models::FineTuning::FineTuningJob::Method::Supervised::Hyperparameters::batch_size? def batch_size=: ( @@ -374,7 +374,7 @@ module OpenAI type batch_size = :auto | Integer module BatchSize - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [:auto, Integer] end @@ -382,7 +382,7 @@ module OpenAI type learning_rate_multiplier = :auto | Float module LearningRateMultiplier - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [:auto, Float] end @@ -390,7 +390,7 @@ module OpenAI type n_epochs = :auto | Integer module NEpochs - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [:auto, Integer] end @@ -400,7 +400,7 @@ module OpenAI type type_ = :supervised | :dpo module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum SUPERVISED: :supervised DPO: :dpo diff --git a/sig/openai/models/fine_tuning/fine_tuning_job_event.rbs b/sig/openai/models/fine_tuning/fine_tuning_job_event.rbs index e9f208ef..dfc567a3 100644 --- a/sig/openai/models/fine_tuning/fine_tuning_job_event.rbs +++ b/sig/openai/models/fine_tuning/fine_tuning_job_event.rbs @@ -14,7 +14,7 @@ module OpenAI type: OpenAI::Models::FineTuning::FineTuningJobEvent::type_ } - class FineTuningJobEvent < OpenAI::BaseModel + class FineTuningJobEvent < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor created_at: Integer @@ -50,7 +50,7 @@ module OpenAI type level = :info | :warn | :error module Level - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum INFO: :info WARN: :warn @@ -62,7 +62,7 @@ module OpenAI type type_ = :message | :metrics module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum MESSAGE: :message METRICS: :metrics diff --git a/sig/openai/models/fine_tuning/fine_tuning_job_wandb_integration.rbs b/sig/openai/models/fine_tuning/fine_tuning_job_wandb_integration.rbs index f5fc5cec..b952d550 100644 --- a/sig/openai/models/fine_tuning/fine_tuning_job_wandb_integration.rbs +++ b/sig/openai/models/fine_tuning/fine_tuning_job_wandb_integration.rbs @@ -11,7 +11,7 @@ module OpenAI tags: ::Array[String] } - class FineTuningJobWandbIntegration < OpenAI::BaseModel + class FineTuningJobWandbIntegration < OpenAI::Internal::Type::BaseModel attr_accessor project: String attr_accessor entity: String? diff --git a/sig/openai/models/fine_tuning/fine_tuning_job_wandb_integration_object.rbs b/sig/openai/models/fine_tuning/fine_tuning_job_wandb_integration_object.rbs index e156cf05..559090c6 100644 --- a/sig/openai/models/fine_tuning/fine_tuning_job_wandb_integration_object.rbs +++ b/sig/openai/models/fine_tuning/fine_tuning_job_wandb_integration_object.rbs @@ -9,7 +9,7 @@ module OpenAI wandb: OpenAI::Models::FineTuning::FineTuningJobWandbIntegration } - class FineTuningJobWandbIntegrationObject < OpenAI::BaseModel + class FineTuningJobWandbIntegrationObject < OpenAI::Internal::Type::BaseModel attr_accessor type: :wandb attr_accessor wandb: OpenAI::Models::FineTuning::FineTuningJobWandbIntegration diff --git a/sig/openai/models/fine_tuning/job_cancel_params.rbs b/sig/openai/models/fine_tuning/job_cancel_params.rbs index 097c08c0..46fadbe0 100644 --- a/sig/openai/models/fine_tuning/job_cancel_params.rbs +++ b/sig/openai/models/fine_tuning/job_cancel_params.rbs @@ -3,7 +3,7 @@ module OpenAI module FineTuning type job_cancel_params = { } & OpenAI::Internal::Type::request_parameters - class JobCancelParams < OpenAI::BaseModel + class JobCancelParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/fine_tuning/job_create_params.rbs b/sig/openai/models/fine_tuning/job_create_params.rbs index fd4f66bb..86773182 100644 --- a/sig/openai/models/fine_tuning/job_create_params.rbs +++ b/sig/openai/models/fine_tuning/job_create_params.rbs @@ -15,7 +15,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class JobCreateParams < OpenAI::BaseModel + class JobCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -68,7 +68,7 @@ module OpenAI | :"gpt-4o-mini" module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, (:"babbage-002" | :"davinci-002" @@ -88,7 +88,7 @@ module OpenAI n_epochs: OpenAI::Models::FineTuning::JobCreateParams::Hyperparameters::n_epochs } - class Hyperparameters < OpenAI::BaseModel + class Hyperparameters < OpenAI::Internal::Type::BaseModel attr_reader batch_size: OpenAI::Models::FineTuning::JobCreateParams::Hyperparameters::batch_size? def batch_size=: ( @@ -118,7 +118,7 @@ module OpenAI type batch_size = :auto | Integer module BatchSize - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [:auto, Integer] end @@ -126,7 +126,7 @@ module OpenAI type learning_rate_multiplier = :auto | Float module LearningRateMultiplier - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [:auto, Float] end @@ -134,7 +134,7 @@ module OpenAI type n_epochs = :auto | Integer module NEpochs - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [:auto, Integer] end @@ -146,7 +146,7 @@ module OpenAI wandb: OpenAI::Models::FineTuning::JobCreateParams::Integration::Wandb } - class Integration < OpenAI::BaseModel + class Integration < OpenAI::Internal::Type::BaseModel attr_accessor type: :wandb attr_accessor wandb: OpenAI::Models::FineTuning::JobCreateParams::Integration::Wandb @@ -166,7 +166,7 @@ module OpenAI tags: ::Array[String] } - class Wandb < OpenAI::BaseModel + class Wandb < OpenAI::Internal::Type::BaseModel attr_accessor project: String attr_accessor entity: String? @@ -195,7 +195,7 @@ module OpenAI type: OpenAI::Models::FineTuning::JobCreateParams::Method::type_ } - class Method < OpenAI::BaseModel + class Method < OpenAI::Internal::Type::BaseModel attr_reader dpo: OpenAI::Models::FineTuning::JobCreateParams::Method::Dpo? def dpo=: ( @@ -227,7 +227,7 @@ module OpenAI hyperparameters: OpenAI::Models::FineTuning::JobCreateParams::Method::Dpo::Hyperparameters } - class Dpo < OpenAI::BaseModel + class Dpo < OpenAI::Internal::Type::BaseModel attr_reader hyperparameters: OpenAI::Models::FineTuning::JobCreateParams::Method::Dpo::Hyperparameters? def hyperparameters=: ( @@ -248,7 +248,7 @@ module OpenAI n_epochs: OpenAI::Models::FineTuning::JobCreateParams::Method::Dpo::Hyperparameters::n_epochs } - class Hyperparameters < OpenAI::BaseModel + class Hyperparameters < OpenAI::Internal::Type::BaseModel attr_reader batch_size: OpenAI::Models::FineTuning::JobCreateParams::Method::Dpo::Hyperparameters::batch_size? def batch_size=: ( @@ -285,7 +285,7 @@ module OpenAI type batch_size = :auto | Integer module BatchSize - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [:auto, Integer] end @@ -293,7 +293,7 @@ module OpenAI type beta = :auto | Float module Beta - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [:auto, Float] end @@ -301,7 +301,7 @@ module OpenAI type learning_rate_multiplier = :auto | Float module LearningRateMultiplier - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [:auto, Float] end @@ -309,7 +309,7 @@ module OpenAI type n_epochs = :auto | Integer module NEpochs - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [:auto, Integer] end @@ -321,7 +321,7 @@ module OpenAI hyperparameters: OpenAI::Models::FineTuning::JobCreateParams::Method::Supervised::Hyperparameters } - class Supervised < OpenAI::BaseModel + class Supervised < OpenAI::Internal::Type::BaseModel attr_reader hyperparameters: OpenAI::Models::FineTuning::JobCreateParams::Method::Supervised::Hyperparameters? def hyperparameters=: ( @@ -341,7 +341,7 @@ module OpenAI n_epochs: OpenAI::Models::FineTuning::JobCreateParams::Method::Supervised::Hyperparameters::n_epochs } - class Hyperparameters < OpenAI::BaseModel + class Hyperparameters < OpenAI::Internal::Type::BaseModel attr_reader batch_size: OpenAI::Models::FineTuning::JobCreateParams::Method::Supervised::Hyperparameters::batch_size? def batch_size=: ( @@ -371,7 +371,7 @@ module OpenAI type batch_size = :auto | Integer module BatchSize - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [:auto, Integer] end @@ -379,7 +379,7 @@ module OpenAI type learning_rate_multiplier = :auto | Float module LearningRateMultiplier - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [:auto, Float] end @@ -387,7 +387,7 @@ module OpenAI type n_epochs = :auto | Integer module NEpochs - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [:auto, Integer] end @@ -397,7 +397,7 @@ module OpenAI type type_ = :supervised | :dpo module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum SUPERVISED: :supervised DPO: :dpo diff --git a/sig/openai/models/fine_tuning/job_list_events_params.rbs b/sig/openai/models/fine_tuning/job_list_events_params.rbs index 2bd10d63..79968e83 100644 --- a/sig/openai/models/fine_tuning/job_list_events_params.rbs +++ b/sig/openai/models/fine_tuning/job_list_events_params.rbs @@ -5,7 +5,7 @@ module OpenAI { after: String, limit: Integer } & OpenAI::Internal::Type::request_parameters - class JobListEventsParams < OpenAI::BaseModel + class JobListEventsParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/fine_tuning/job_list_params.rbs b/sig/openai/models/fine_tuning/job_list_params.rbs index a09f5864..b32ac587 100644 --- a/sig/openai/models/fine_tuning/job_list_params.rbs +++ b/sig/openai/models/fine_tuning/job_list_params.rbs @@ -5,7 +5,7 @@ module OpenAI { after: String, limit: Integer, metadata: ::Hash[Symbol, String]? } & OpenAI::Internal::Type::request_parameters - class JobListParams < OpenAI::BaseModel + class JobListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/fine_tuning/job_retrieve_params.rbs b/sig/openai/models/fine_tuning/job_retrieve_params.rbs index 466850ae..179216b1 100644 --- a/sig/openai/models/fine_tuning/job_retrieve_params.rbs +++ b/sig/openai/models/fine_tuning/job_retrieve_params.rbs @@ -4,7 +4,7 @@ module OpenAI type job_retrieve_params = { } & OpenAI::Internal::Type::request_parameters - class JobRetrieveParams < OpenAI::BaseModel + class JobRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/fine_tuning/jobs/checkpoint_list_params.rbs b/sig/openai/models/fine_tuning/jobs/checkpoint_list_params.rbs index dc5aa5a0..012833d8 100644 --- a/sig/openai/models/fine_tuning/jobs/checkpoint_list_params.rbs +++ b/sig/openai/models/fine_tuning/jobs/checkpoint_list_params.rbs @@ -6,7 +6,7 @@ module OpenAI { after: String, limit: Integer } & OpenAI::Internal::Type::request_parameters - class CheckpointListParams < OpenAI::BaseModel + class CheckpointListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/fine_tuning/jobs/fine_tuning_job_checkpoint.rbs b/sig/openai/models/fine_tuning/jobs/fine_tuning_job_checkpoint.rbs index 4f7cf1b7..1f11cb97 100644 --- a/sig/openai/models/fine_tuning/jobs/fine_tuning_job_checkpoint.rbs +++ b/sig/openai/models/fine_tuning/jobs/fine_tuning_job_checkpoint.rbs @@ -13,7 +13,7 @@ module OpenAI step_number: Integer } - class FineTuningJobCheckpoint < OpenAI::BaseModel + class FineTuningJobCheckpoint < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor created_at: Integer @@ -51,7 +51,7 @@ module OpenAI valid_mean_token_accuracy: Float } - class Metrics < OpenAI::BaseModel + class Metrics < OpenAI::Internal::Type::BaseModel attr_reader full_valid_loss: Float? def full_valid_loss=: (Float) -> Float diff --git a/sig/openai/models/function_definition.rbs b/sig/openai/models/function_definition.rbs index 194ec56f..c8acb986 100644 --- a/sig/openai/models/function_definition.rbs +++ b/sig/openai/models/function_definition.rbs @@ -8,7 +8,7 @@ module OpenAI strict: bool? } - class FunctionDefinition < OpenAI::BaseModel + class FunctionDefinition < OpenAI::Internal::Type::BaseModel attr_accessor name: String attr_reader description: String? diff --git a/sig/openai/models/image.rbs b/sig/openai/models/image.rbs index ed120c3b..f95aa74f 100644 --- a/sig/openai/models/image.rbs +++ b/sig/openai/models/image.rbs @@ -2,7 +2,7 @@ module OpenAI module Models type image = { :b64_json => String, revised_prompt: String, url: String } - class Image < OpenAI::BaseModel + class Image < OpenAI::Internal::Type::BaseModel attr_reader b64_json: String? def b64_json=: (String) -> String diff --git a/sig/openai/models/image_create_variation_params.rbs b/sig/openai/models/image_create_variation_params.rbs index ebbc6451..7b3071d3 100644 --- a/sig/openai/models/image_create_variation_params.rbs +++ b/sig/openai/models/image_create_variation_params.rbs @@ -11,7 +11,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class ImageCreateVariationParams < OpenAI::BaseModel + class ImageCreateVariationParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -44,7 +44,7 @@ module OpenAI type model = String | OpenAI::Models::image_model module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, OpenAI::Models::image_model] end @@ -52,7 +52,7 @@ module OpenAI type response_format = :url | :b64_json module ResponseFormat - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum URL: :url B64_JSON: :b64_json @@ -63,7 +63,7 @@ module OpenAI type size = :"256x256" | :"512x512" | :"1024x1024" module Size - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum SIZE_256X256: :"256x256" SIZE_512X512: :"512x512" diff --git a/sig/openai/models/image_edit_params.rbs b/sig/openai/models/image_edit_params.rbs index 616d4a05..9cb884a7 100644 --- a/sig/openai/models/image_edit_params.rbs +++ b/sig/openai/models/image_edit_params.rbs @@ -13,7 +13,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class ImageEditParams < OpenAI::BaseModel + class ImageEditParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -54,7 +54,7 @@ module OpenAI type model = String | OpenAI::Models::image_model module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, OpenAI::Models::image_model] end @@ -62,7 +62,7 @@ module OpenAI type response_format = :url | :b64_json module ResponseFormat - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum URL: :url B64_JSON: :b64_json @@ -73,7 +73,7 @@ module OpenAI type size = :"256x256" | :"512x512" | :"1024x1024" module Size - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum SIZE_256X256: :"256x256" SIZE_512X512: :"512x512" diff --git a/sig/openai/models/image_generate_params.rbs b/sig/openai/models/image_generate_params.rbs index 5af7ffe3..869a9475 100644 --- a/sig/openai/models/image_generate_params.rbs +++ b/sig/openai/models/image_generate_params.rbs @@ -13,7 +13,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class ImageGenerateParams < OpenAI::BaseModel + class ImageGenerateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -56,7 +56,7 @@ module OpenAI type model = String | OpenAI::Models::image_model module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, OpenAI::Models::image_model] end @@ -64,7 +64,7 @@ module OpenAI type quality = :standard | :hd module Quality - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum STANDARD: :standard HD: :hd @@ -75,7 +75,7 @@ module OpenAI type response_format = :url | :b64_json module ResponseFormat - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum URL: :url B64_JSON: :b64_json @@ -87,7 +87,7 @@ module OpenAI :"256x256" | :"512x512" | :"1024x1024" | :"1792x1024" | :"1024x1792" module Size - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum SIZE_256X256: :"256x256" SIZE_512X512: :"512x512" @@ -101,7 +101,7 @@ module OpenAI type style = :vivid | :natural module Style - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum VIVID: :vivid NATURAL: :natural diff --git a/sig/openai/models/image_model.rbs b/sig/openai/models/image_model.rbs index 97e2e64f..31f6927a 100644 --- a/sig/openai/models/image_model.rbs +++ b/sig/openai/models/image_model.rbs @@ -3,7 +3,7 @@ module OpenAI type image_model = :"dall-e-2" | :"dall-e-3" module ImageModel - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum DALL_E_2: :"dall-e-2" DALL_E_3: :"dall-e-3" diff --git a/sig/openai/models/images_response.rbs b/sig/openai/models/images_response.rbs index cafb98c6..e5608584 100644 --- a/sig/openai/models/images_response.rbs +++ b/sig/openai/models/images_response.rbs @@ -3,7 +3,7 @@ module OpenAI type images_response = { created: Integer, data: ::Array[OpenAI::Models::Image] } - class ImagesResponse < OpenAI::BaseModel + class ImagesResponse < OpenAI::Internal::Type::BaseModel attr_accessor created: Integer attr_accessor data: ::Array[OpenAI::Models::Image] diff --git a/sig/openai/models/model.rbs b/sig/openai/models/model.rbs index 3540d0d1..88000ee1 100644 --- a/sig/openai/models/model.rbs +++ b/sig/openai/models/model.rbs @@ -3,7 +3,7 @@ module OpenAI type model = { id: String, created: Integer, object: :model, owned_by: String } - class Model < OpenAI::BaseModel + class Model < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor created: Integer diff --git a/sig/openai/models/model_delete_params.rbs b/sig/openai/models/model_delete_params.rbs index 2f5cfd40..ab28b6e7 100644 --- a/sig/openai/models/model_delete_params.rbs +++ b/sig/openai/models/model_delete_params.rbs @@ -2,7 +2,7 @@ module OpenAI module Models type model_delete_params = { } & OpenAI::Internal::Type::request_parameters - class ModelDeleteParams < OpenAI::BaseModel + class ModelDeleteParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/model_deleted.rbs b/sig/openai/models/model_deleted.rbs index 6297c483..f1c1b5f9 100644 --- a/sig/openai/models/model_deleted.rbs +++ b/sig/openai/models/model_deleted.rbs @@ -2,7 +2,7 @@ module OpenAI module Models type model_deleted = { id: String, deleted: bool, object: String } - class ModelDeleted < OpenAI::BaseModel + class ModelDeleted < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor deleted: bool diff --git a/sig/openai/models/model_list_params.rbs b/sig/openai/models/model_list_params.rbs index 4511e755..ffbe02ed 100644 --- a/sig/openai/models/model_list_params.rbs +++ b/sig/openai/models/model_list_params.rbs @@ -2,7 +2,7 @@ module OpenAI module Models type model_list_params = { } & OpenAI::Internal::Type::request_parameters - class ModelListParams < OpenAI::BaseModel + class ModelListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/model_retrieve_params.rbs b/sig/openai/models/model_retrieve_params.rbs index dfcd6daa..8cf3afab 100644 --- a/sig/openai/models/model_retrieve_params.rbs +++ b/sig/openai/models/model_retrieve_params.rbs @@ -3,7 +3,7 @@ module OpenAI type model_retrieve_params = { } & OpenAI::Internal::Type::request_parameters - class ModelRetrieveParams < OpenAI::BaseModel + class ModelRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/moderation.rbs b/sig/openai/models/moderation.rbs index 9967761f..2a38be49 100644 --- a/sig/openai/models/moderation.rbs +++ b/sig/openai/models/moderation.rbs @@ -8,7 +8,7 @@ module OpenAI flagged: bool } - class Moderation < OpenAI::BaseModel + class Moderation < OpenAI::Internal::Type::BaseModel attr_accessor categories: OpenAI::Models::Moderation::Categories attr_accessor category_applied_input_types: OpenAI::Models::Moderation::CategoryAppliedInputTypes @@ -43,7 +43,7 @@ module OpenAI violence_graphic: bool } - class Categories < OpenAI::BaseModel + class Categories < OpenAI::Internal::Type::BaseModel attr_accessor harassment: bool attr_accessor harassment_threatening: bool @@ -106,7 +106,7 @@ module OpenAI violence_graphic: ::Array[OpenAI::Models::Moderation::CategoryAppliedInputTypes::violence_graphic] } - class CategoryAppliedInputTypes < OpenAI::BaseModel + class CategoryAppliedInputTypes < OpenAI::Internal::Type::BaseModel attr_accessor harassment: ::Array[OpenAI::Models::Moderation::CategoryAppliedInputTypes::harassment] attr_accessor harassment_threatening: ::Array[OpenAI::Models::Moderation::CategoryAppliedInputTypes::harassment_threatening] @@ -154,7 +154,7 @@ module OpenAI type harassment = :text module Harassment - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT: :text @@ -164,7 +164,7 @@ module OpenAI type harassment_threatening = :text module HarassmentThreatening - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT: :text @@ -174,7 +174,7 @@ module OpenAI type hate = :text module Hate - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT: :text @@ -184,7 +184,7 @@ module OpenAI type hate_threatening = :text module HateThreatening - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT: :text @@ -194,7 +194,7 @@ module OpenAI type illicit = :text module Illicit - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT: :text @@ -204,7 +204,7 @@ module OpenAI type illicit_violent = :text module IllicitViolent - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT: :text @@ -214,7 +214,7 @@ module OpenAI type self_harm = :text | :image module SelfHarm - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT: :text IMAGE: :image @@ -225,7 +225,7 @@ module OpenAI type self_harm_instruction = :text | :image module SelfHarmInstruction - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT: :text IMAGE: :image @@ -236,7 +236,7 @@ module OpenAI type self_harm_intent = :text | :image module SelfHarmIntent - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT: :text IMAGE: :image @@ -247,7 +247,7 @@ module OpenAI type sexual = :text | :image module Sexual - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT: :text IMAGE: :image @@ -258,7 +258,7 @@ module OpenAI type sexual_minor = :text module SexualMinor - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT: :text @@ -268,7 +268,7 @@ module OpenAI type violence = :text | :image module Violence - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT: :text IMAGE: :image @@ -279,7 +279,7 @@ module OpenAI type violence_graphic = :text | :image module ViolenceGraphic - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT: :text IMAGE: :image @@ -305,7 +305,7 @@ module OpenAI violence_graphic: Float } - class CategoryScores < OpenAI::BaseModel + class CategoryScores < OpenAI::Internal::Type::BaseModel attr_accessor harassment: Float attr_accessor harassment_threatening: Float diff --git a/sig/openai/models/moderation_create_params.rbs b/sig/openai/models/moderation_create_params.rbs index 833c5ff4..e8c1e576 100644 --- a/sig/openai/models/moderation_create_params.rbs +++ b/sig/openai/models/moderation_create_params.rbs @@ -7,7 +7,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class ModerationCreateParams < OpenAI::BaseModel + class ModerationCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -33,7 +33,7 @@ module OpenAI | ::Array[OpenAI::Models::moderation_multi_modal_input] module Input - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, ::Array[String], ::Array[OpenAI::Models::moderation_multi_modal_input]] @@ -50,7 +50,7 @@ module OpenAI type model = String | OpenAI::Models::moderation_model module Model - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, OpenAI::Models::moderation_model] end diff --git a/sig/openai/models/moderation_create_response.rbs b/sig/openai/models/moderation_create_response.rbs index 5934337f..d6432898 100644 --- a/sig/openai/models/moderation_create_response.rbs +++ b/sig/openai/models/moderation_create_response.rbs @@ -7,7 +7,7 @@ module OpenAI results: ::Array[OpenAI::Models::Moderation] } - class ModerationCreateResponse < OpenAI::BaseModel + class ModerationCreateResponse < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor model: String diff --git a/sig/openai/models/moderation_image_url_input.rbs b/sig/openai/models/moderation_image_url_input.rbs index e23f649d..4aa41a71 100644 --- a/sig/openai/models/moderation_image_url_input.rbs +++ b/sig/openai/models/moderation_image_url_input.rbs @@ -6,7 +6,7 @@ module OpenAI type: :image_url } - class ModerationImageURLInput < OpenAI::BaseModel + class ModerationImageURLInput < OpenAI::Internal::Type::BaseModel attr_accessor image_url: OpenAI::Models::ModerationImageURLInput::ImageURL attr_accessor type: :image_url @@ -20,7 +20,7 @@ module OpenAI type image_url = { url: String } - class ImageURL < OpenAI::BaseModel + class ImageURL < OpenAI::Internal::Type::BaseModel attr_accessor url: String def initialize: (url: String) -> void diff --git a/sig/openai/models/moderation_model.rbs b/sig/openai/models/moderation_model.rbs index 1f6ab3fd..228342be 100644 --- a/sig/openai/models/moderation_model.rbs +++ b/sig/openai/models/moderation_model.rbs @@ -7,7 +7,7 @@ module OpenAI | :"text-moderation-stable" module ModerationModel - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum OMNI_MODERATION_LATEST: :"omni-moderation-latest" OMNI_MODERATION_2024_09_26: :"omni-moderation-2024-09-26" diff --git a/sig/openai/models/moderation_multi_modal_input.rbs b/sig/openai/models/moderation_multi_modal_input.rbs index 967234fc..61d672c1 100644 --- a/sig/openai/models/moderation_multi_modal_input.rbs +++ b/sig/openai/models/moderation_multi_modal_input.rbs @@ -5,7 +5,7 @@ module OpenAI | OpenAI::Models::ModerationTextInput module ModerationMultiModalInput - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::ModerationImageURLInput, OpenAI::Models::ModerationTextInput] end diff --git a/sig/openai/models/moderation_text_input.rbs b/sig/openai/models/moderation_text_input.rbs index 45143b65..c6d70b96 100644 --- a/sig/openai/models/moderation_text_input.rbs +++ b/sig/openai/models/moderation_text_input.rbs @@ -2,7 +2,7 @@ module OpenAI module Models type moderation_text_input = { text: String, type: :text } - class ModerationTextInput < OpenAI::BaseModel + class ModerationTextInput < OpenAI::Internal::Type::BaseModel attr_accessor text: String attr_accessor type: :text diff --git a/sig/openai/models/other_file_chunking_strategy_object.rbs b/sig/openai/models/other_file_chunking_strategy_object.rbs index 60c24b65..01290f70 100644 --- a/sig/openai/models/other_file_chunking_strategy_object.rbs +++ b/sig/openai/models/other_file_chunking_strategy_object.rbs @@ -2,7 +2,7 @@ module OpenAI module Models type other_file_chunking_strategy_object = { type: :other } - class OtherFileChunkingStrategyObject < OpenAI::BaseModel + class OtherFileChunkingStrategyObject < OpenAI::Internal::Type::BaseModel attr_accessor type: :other def initialize: (?type: :other) -> void diff --git a/sig/openai/models/reasoning.rbs b/sig/openai/models/reasoning.rbs index d4060722..c299e779 100644 --- a/sig/openai/models/reasoning.rbs +++ b/sig/openai/models/reasoning.rbs @@ -6,7 +6,7 @@ module OpenAI generate_summary: OpenAI::Models::Reasoning::generate_summary? } - class Reasoning < OpenAI::BaseModel + class Reasoning < OpenAI::Internal::Type::BaseModel attr_accessor effort: OpenAI::Models::reasoning_effort? attr_accessor generate_summary: OpenAI::Models::Reasoning::generate_summary? @@ -21,7 +21,7 @@ module OpenAI type generate_summary = :concise | :detailed module GenerateSummary - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum CONCISE: :concise DETAILED: :detailed diff --git a/sig/openai/models/reasoning_effort.rbs b/sig/openai/models/reasoning_effort.rbs index 7ee7ea11..d449d634 100644 --- a/sig/openai/models/reasoning_effort.rbs +++ b/sig/openai/models/reasoning_effort.rbs @@ -3,7 +3,7 @@ module OpenAI type reasoning_effort = :low | :medium | :high module ReasoningEffort - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum LOW: :low MEDIUM: :medium diff --git a/sig/openai/models/response_format_json_object.rbs b/sig/openai/models/response_format_json_object.rbs index f228d234..09c67374 100644 --- a/sig/openai/models/response_format_json_object.rbs +++ b/sig/openai/models/response_format_json_object.rbs @@ -2,7 +2,7 @@ module OpenAI module Models type response_format_json_object = { type: :json_object } - class ResponseFormatJSONObject < OpenAI::BaseModel + class ResponseFormatJSONObject < OpenAI::Internal::Type::BaseModel attr_accessor type: :json_object def initialize: (?type: :json_object) -> void diff --git a/sig/openai/models/response_format_json_schema.rbs b/sig/openai/models/response_format_json_schema.rbs index 05e4a025..4d9f4d87 100644 --- a/sig/openai/models/response_format_json_schema.rbs +++ b/sig/openai/models/response_format_json_schema.rbs @@ -6,7 +6,7 @@ module OpenAI type: :json_schema } - class ResponseFormatJSONSchema < OpenAI::BaseModel + class ResponseFormatJSONSchema < OpenAI::Internal::Type::BaseModel attr_accessor json_schema: OpenAI::Models::ResponseFormatJSONSchema::JSONSchema attr_accessor type: :json_schema @@ -26,7 +26,7 @@ module OpenAI strict: bool? } - class JSONSchema < OpenAI::BaseModel + class JSONSchema < OpenAI::Internal::Type::BaseModel attr_accessor name: String attr_reader description: String? diff --git a/sig/openai/models/response_format_text.rbs b/sig/openai/models/response_format_text.rbs index 691bace2..7b892f07 100644 --- a/sig/openai/models/response_format_text.rbs +++ b/sig/openai/models/response_format_text.rbs @@ -2,7 +2,7 @@ module OpenAI module Models type response_format_text = { type: :text } - class ResponseFormatText < OpenAI::BaseModel + class ResponseFormatText < OpenAI::Internal::Type::BaseModel attr_accessor type: :text def initialize: (?type: :text) -> void diff --git a/sig/openai/models/responses/computer_tool.rbs b/sig/openai/models/responses/computer_tool.rbs index 0b984fde..f30d8909 100644 --- a/sig/openai/models/responses/computer_tool.rbs +++ b/sig/openai/models/responses/computer_tool.rbs @@ -9,7 +9,7 @@ module OpenAI type: :computer_use_preview } - class ComputerTool < OpenAI::BaseModel + class ComputerTool < OpenAI::Internal::Type::BaseModel attr_accessor display_height: Float attr_accessor display_width: Float @@ -30,7 +30,7 @@ module OpenAI type environment = :mac | :windows | :ubuntu | :browser module Environment - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum MAC: :mac WINDOWS: :windows diff --git a/sig/openai/models/responses/easy_input_message.rbs b/sig/openai/models/responses/easy_input_message.rbs index 04936217..1120ea2c 100644 --- a/sig/openai/models/responses/easy_input_message.rbs +++ b/sig/openai/models/responses/easy_input_message.rbs @@ -8,7 +8,7 @@ module OpenAI type: OpenAI::Models::Responses::EasyInputMessage::type_ } - class EasyInputMessage < OpenAI::BaseModel + class EasyInputMessage < OpenAI::Internal::Type::BaseModel attr_accessor content: OpenAI::Models::Responses::EasyInputMessage::content attr_accessor role: OpenAI::Models::Responses::EasyInputMessage::role @@ -32,7 +32,7 @@ module OpenAI | OpenAI::Models::Responses::response_input_message_content_list module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, OpenAI::Models::Responses::response_input_message_content_list] end @@ -40,7 +40,7 @@ module OpenAI type role = :user | :assistant | :system | :developer module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum USER: :user ASSISTANT: :assistant @@ -53,7 +53,7 @@ module OpenAI type type_ = :message module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum MESSAGE: :message diff --git a/sig/openai/models/responses/file_search_tool.rbs b/sig/openai/models/responses/file_search_tool.rbs index 69476cbd..6c174053 100644 --- a/sig/openai/models/responses/file_search_tool.rbs +++ b/sig/openai/models/responses/file_search_tool.rbs @@ -10,7 +10,7 @@ module OpenAI ranking_options: OpenAI::Models::Responses::FileSearchTool::RankingOptions } - class FileSearchTool < OpenAI::BaseModel + class FileSearchTool < OpenAI::Internal::Type::BaseModel attr_accessor type: :file_search attr_accessor vector_store_ids: ::Array[String] @@ -45,7 +45,7 @@ module OpenAI OpenAI::Models::ComparisonFilter | OpenAI::Models::CompoundFilter module Filters - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::ComparisonFilter, OpenAI::Models::CompoundFilter] end @@ -56,7 +56,7 @@ module OpenAI score_threshold: Float } - class RankingOptions < OpenAI::BaseModel + class RankingOptions < OpenAI::Internal::Type::BaseModel attr_reader ranker: OpenAI::Models::Responses::FileSearchTool::RankingOptions::ranker? def ranker=: ( @@ -77,7 +77,7 @@ module OpenAI type ranker = :auto | :"default-2024-11-15" module Ranker - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO: :auto DEFAULT_2024_11_15: :"default-2024-11-15" diff --git a/sig/openai/models/responses/function_tool.rbs b/sig/openai/models/responses/function_tool.rbs index b0e5aa03..f69630ac 100644 --- a/sig/openai/models/responses/function_tool.rbs +++ b/sig/openai/models/responses/function_tool.rbs @@ -10,7 +10,7 @@ module OpenAI description: String? } - class FunctionTool < OpenAI::BaseModel + class FunctionTool < OpenAI::Internal::Type::BaseModel attr_accessor name: String attr_accessor parameters: ::Hash[Symbol, top] diff --git a/sig/openai/models/responses/input_item_list_params.rbs b/sig/openai/models/responses/input_item_list_params.rbs index 18327460..3e047f28 100644 --- a/sig/openai/models/responses/input_item_list_params.rbs +++ b/sig/openai/models/responses/input_item_list_params.rbs @@ -11,7 +11,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class InputItemListParams < OpenAI::BaseModel + class InputItemListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -53,7 +53,7 @@ module OpenAI type order = :asc | :desc module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASC: :asc DESC: :desc diff --git a/sig/openai/models/responses/response.rbs b/sig/openai/models/responses/response.rbs index 254aa7e1..86acfc4c 100644 --- a/sig/openai/models/responses/response.rbs +++ b/sig/openai/models/responses/response.rbs @@ -27,7 +27,7 @@ module OpenAI user: String } - class Response < OpenAI::BaseModel + class Response < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor created_at: Float @@ -118,7 +118,7 @@ module OpenAI reason: OpenAI::Models::Responses::Response::IncompleteDetails::reason } - class IncompleteDetails < OpenAI::BaseModel + class IncompleteDetails < OpenAI::Internal::Type::BaseModel attr_reader reason: OpenAI::Models::Responses::Response::IncompleteDetails::reason? def reason=: ( @@ -134,7 +134,7 @@ module OpenAI type reason = :max_output_tokens | :content_filter module Reason - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum MAX_OUTPUT_TOKENS: :max_output_tokens CONTENT_FILTER: :content_filter @@ -149,7 +149,7 @@ module OpenAI | OpenAI::Models::Responses::ToolChoiceFunction module ToolChoice - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Responses::tool_choice_options, OpenAI::Models::Responses::ToolChoiceTypes, OpenAI::Models::Responses::ToolChoiceFunction] end @@ -157,7 +157,7 @@ module OpenAI type truncation = :auto | :disabled module Truncation - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO: :auto DISABLED: :disabled diff --git a/sig/openai/models/responses/response_audio_delta_event.rbs b/sig/openai/models/responses/response_audio_delta_event.rbs index f6a9c6fd..7a218696 100644 --- a/sig/openai/models/responses/response_audio_delta_event.rbs +++ b/sig/openai/models/responses/response_audio_delta_event.rbs @@ -4,7 +4,7 @@ module OpenAI type response_audio_delta_event = { delta: String, type: :"response.audio.delta" } - class ResponseAudioDeltaEvent < OpenAI::BaseModel + class ResponseAudioDeltaEvent < OpenAI::Internal::Type::BaseModel attr_accessor delta: String attr_accessor type: :"response.audio.delta" diff --git a/sig/openai/models/responses/response_audio_done_event.rbs b/sig/openai/models/responses/response_audio_done_event.rbs index 9449f329..cbff6d81 100644 --- a/sig/openai/models/responses/response_audio_done_event.rbs +++ b/sig/openai/models/responses/response_audio_done_event.rbs @@ -3,7 +3,7 @@ module OpenAI module Responses type response_audio_done_event = { type: :"response.audio.done" } - class ResponseAudioDoneEvent < OpenAI::BaseModel + class ResponseAudioDoneEvent < OpenAI::Internal::Type::BaseModel attr_accessor type: :"response.audio.done" def initialize: (?type: :"response.audio.done") -> void diff --git a/sig/openai/models/responses/response_audio_transcript_delta_event.rbs b/sig/openai/models/responses/response_audio_transcript_delta_event.rbs index b1e7a534..6f8fc8cb 100644 --- a/sig/openai/models/responses/response_audio_transcript_delta_event.rbs +++ b/sig/openai/models/responses/response_audio_transcript_delta_event.rbs @@ -4,7 +4,7 @@ module OpenAI type response_audio_transcript_delta_event = { delta: String, type: :"response.audio.transcript.delta" } - class ResponseAudioTranscriptDeltaEvent < OpenAI::BaseModel + class ResponseAudioTranscriptDeltaEvent < OpenAI::Internal::Type::BaseModel attr_accessor delta: String attr_accessor type: :"response.audio.transcript.delta" diff --git a/sig/openai/models/responses/response_audio_transcript_done_event.rbs b/sig/openai/models/responses/response_audio_transcript_done_event.rbs index ca5e5241..9d7072d8 100644 --- a/sig/openai/models/responses/response_audio_transcript_done_event.rbs +++ b/sig/openai/models/responses/response_audio_transcript_done_event.rbs @@ -4,7 +4,7 @@ module OpenAI type response_audio_transcript_done_event = { type: :"response.audio.transcript.done" } - class ResponseAudioTranscriptDoneEvent < OpenAI::BaseModel + class ResponseAudioTranscriptDoneEvent < OpenAI::Internal::Type::BaseModel attr_accessor type: :"response.audio.transcript.done" def initialize: (?type: :"response.audio.transcript.done") -> void diff --git a/sig/openai/models/responses/response_code_interpreter_call_code_delta_event.rbs b/sig/openai/models/responses/response_code_interpreter_call_code_delta_event.rbs index a330e86c..23837b29 100644 --- a/sig/openai/models/responses/response_code_interpreter_call_code_delta_event.rbs +++ b/sig/openai/models/responses/response_code_interpreter_call_code_delta_event.rbs @@ -8,7 +8,7 @@ module OpenAI type: :"response.code_interpreter_call.code.delta" } - class ResponseCodeInterpreterCallCodeDeltaEvent < OpenAI::BaseModel + class ResponseCodeInterpreterCallCodeDeltaEvent < OpenAI::Internal::Type::BaseModel attr_accessor delta: String attr_accessor output_index: Integer diff --git a/sig/openai/models/responses/response_code_interpreter_call_code_done_event.rbs b/sig/openai/models/responses/response_code_interpreter_call_code_done_event.rbs index 45e0a1fd..d1358071 100644 --- a/sig/openai/models/responses/response_code_interpreter_call_code_done_event.rbs +++ b/sig/openai/models/responses/response_code_interpreter_call_code_done_event.rbs @@ -8,7 +8,7 @@ module OpenAI type: :"response.code_interpreter_call.code.done" } - class ResponseCodeInterpreterCallCodeDoneEvent < OpenAI::BaseModel + class ResponseCodeInterpreterCallCodeDoneEvent < OpenAI::Internal::Type::BaseModel attr_accessor code: String attr_accessor output_index: Integer diff --git a/sig/openai/models/responses/response_code_interpreter_call_completed_event.rbs b/sig/openai/models/responses/response_code_interpreter_call_completed_event.rbs index 46aae17c..840dabb1 100644 --- a/sig/openai/models/responses/response_code_interpreter_call_completed_event.rbs +++ b/sig/openai/models/responses/response_code_interpreter_call_completed_event.rbs @@ -8,7 +8,7 @@ module OpenAI type: :"response.code_interpreter_call.completed" } - class ResponseCodeInterpreterCallCompletedEvent < OpenAI::BaseModel + class ResponseCodeInterpreterCallCompletedEvent < OpenAI::Internal::Type::BaseModel attr_accessor code_interpreter_call: OpenAI::Models::Responses::ResponseCodeInterpreterToolCall attr_accessor output_index: Integer diff --git a/sig/openai/models/responses/response_code_interpreter_call_in_progress_event.rbs b/sig/openai/models/responses/response_code_interpreter_call_in_progress_event.rbs index a5384a8c..f19814d0 100644 --- a/sig/openai/models/responses/response_code_interpreter_call_in_progress_event.rbs +++ b/sig/openai/models/responses/response_code_interpreter_call_in_progress_event.rbs @@ -8,7 +8,7 @@ module OpenAI type: :"response.code_interpreter_call.in_progress" } - class ResponseCodeInterpreterCallInProgressEvent < OpenAI::BaseModel + class ResponseCodeInterpreterCallInProgressEvent < OpenAI::Internal::Type::BaseModel attr_accessor code_interpreter_call: OpenAI::Models::Responses::ResponseCodeInterpreterToolCall attr_accessor output_index: Integer diff --git a/sig/openai/models/responses/response_code_interpreter_call_interpreting_event.rbs b/sig/openai/models/responses/response_code_interpreter_call_interpreting_event.rbs index 3fcff8bf..fdaebc1e 100644 --- a/sig/openai/models/responses/response_code_interpreter_call_interpreting_event.rbs +++ b/sig/openai/models/responses/response_code_interpreter_call_interpreting_event.rbs @@ -8,7 +8,7 @@ module OpenAI type: :"response.code_interpreter_call.interpreting" } - class ResponseCodeInterpreterCallInterpretingEvent < OpenAI::BaseModel + class ResponseCodeInterpreterCallInterpretingEvent < OpenAI::Internal::Type::BaseModel attr_accessor code_interpreter_call: OpenAI::Models::Responses::ResponseCodeInterpreterToolCall attr_accessor output_index: Integer diff --git a/sig/openai/models/responses/response_code_interpreter_tool_call.rbs b/sig/openai/models/responses/response_code_interpreter_tool_call.rbs index 0a64a853..b6aa4db6 100644 --- a/sig/openai/models/responses/response_code_interpreter_tool_call.rbs +++ b/sig/openai/models/responses/response_code_interpreter_tool_call.rbs @@ -10,7 +10,7 @@ module OpenAI type: :code_interpreter_call } - class ResponseCodeInterpreterToolCall < OpenAI::BaseModel + class ResponseCodeInterpreterToolCall < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor code: String @@ -36,11 +36,11 @@ module OpenAI | OpenAI::Models::Responses::ResponseCodeInterpreterToolCall::Result::Files module Result - extend OpenAI::Union + extend OpenAI::Internal::Type::Union type logs = { logs: String, type: :logs } - class Logs < OpenAI::BaseModel + class Logs < OpenAI::Internal::Type::BaseModel attr_accessor logs: String attr_accessor type: :logs @@ -56,7 +56,7 @@ module OpenAI type: :files } - class Files < OpenAI::BaseModel + class Files < OpenAI::Internal::Type::BaseModel attr_accessor files: ::Array[OpenAI::Models::Responses::ResponseCodeInterpreterToolCall::Result::Files::File] attr_accessor type: :files @@ -70,7 +70,7 @@ module OpenAI type file = { file_id: String, mime_type: String } - class File < OpenAI::BaseModel + class File < OpenAI::Internal::Type::BaseModel attr_accessor file_id: String attr_accessor mime_type: String @@ -87,7 +87,7 @@ module OpenAI type status = :in_progress | :interpreting | :completed module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS: :in_progress INTERPRETING: :interpreting diff --git a/sig/openai/models/responses/response_completed_event.rbs b/sig/openai/models/responses/response_completed_event.rbs index 372c8eb9..5ef101ac 100644 --- a/sig/openai/models/responses/response_completed_event.rbs +++ b/sig/openai/models/responses/response_completed_event.rbs @@ -7,7 +7,7 @@ module OpenAI type: :"response.completed" } - class ResponseCompletedEvent < OpenAI::BaseModel + class ResponseCompletedEvent < OpenAI::Internal::Type::BaseModel attr_accessor response: OpenAI::Models::Responses::Response attr_accessor type: :"response.completed" diff --git a/sig/openai/models/responses/response_computer_tool_call.rbs b/sig/openai/models/responses/response_computer_tool_call.rbs index 1cf4a066..ab81a03e 100644 --- a/sig/openai/models/responses/response_computer_tool_call.rbs +++ b/sig/openai/models/responses/response_computer_tool_call.rbs @@ -11,7 +11,7 @@ module OpenAI type: OpenAI::Models::Responses::ResponseComputerToolCall::type_ } - class ResponseComputerToolCall < OpenAI::BaseModel + class ResponseComputerToolCall < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor action: OpenAI::Models::Responses::ResponseComputerToolCall::action @@ -47,7 +47,7 @@ module OpenAI | OpenAI::Models::Responses::ResponseComputerToolCall::Action::Wait module Action - extend OpenAI::Union + extend OpenAI::Internal::Type::Union type click = { @@ -57,7 +57,7 @@ module OpenAI y_: Integer } - class Click < OpenAI::BaseModel + class Click < OpenAI::Internal::Type::BaseModel attr_accessor button: OpenAI::Models::Responses::ResponseComputerToolCall::Action::Click::button attr_accessor type: :click @@ -78,7 +78,7 @@ module OpenAI type button = :left | :right | :wheel | :back | :forward module Button - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum LEFT: :left RIGHT: :right @@ -92,7 +92,7 @@ module OpenAI type double_click = { type: :double_click, x: Integer, y_: Integer } - class DoubleClick < OpenAI::BaseModel + class DoubleClick < OpenAI::Internal::Type::BaseModel attr_accessor type: :double_click attr_accessor x: Integer @@ -114,7 +114,7 @@ module OpenAI type: :drag } - class Drag < OpenAI::BaseModel + class Drag < OpenAI::Internal::Type::BaseModel attr_accessor path: ::Array[OpenAI::Models::Responses::ResponseComputerToolCall::Action::Drag::Path] attr_accessor type: :drag @@ -128,7 +128,7 @@ module OpenAI type path = { x: Integer, y_: Integer } - class Path < OpenAI::BaseModel + class Path < OpenAI::Internal::Type::BaseModel attr_accessor x: Integer attr_accessor y_: Integer @@ -141,7 +141,7 @@ module OpenAI type keypress = { keys: ::Array[String], type: :keypress } - class Keypress < OpenAI::BaseModel + class Keypress < OpenAI::Internal::Type::BaseModel attr_accessor keys: ::Array[String] attr_accessor type: :keypress @@ -153,7 +153,7 @@ module OpenAI type move = { type: :move, x: Integer, y_: Integer } - class Move < OpenAI::BaseModel + class Move < OpenAI::Internal::Type::BaseModel attr_accessor type: :move attr_accessor x: Integer @@ -167,7 +167,7 @@ module OpenAI type screenshot = { type: :screenshot } - class Screenshot < OpenAI::BaseModel + class Screenshot < OpenAI::Internal::Type::BaseModel attr_accessor type: :screenshot def initialize: (?type: :screenshot) -> void @@ -184,7 +184,7 @@ module OpenAI y_: Integer } - class Scroll < OpenAI::BaseModel + class Scroll < OpenAI::Internal::Type::BaseModel attr_accessor scroll_x: Integer attr_accessor scroll_y: Integer @@ -208,7 +208,7 @@ module OpenAI type type_ = { text: String, type: :type } - class Type < OpenAI::BaseModel + class Type < OpenAI::Internal::Type::BaseModel attr_accessor text: String attr_accessor type: :type @@ -220,7 +220,7 @@ module OpenAI type wait = { type: :wait } - class Wait < OpenAI::BaseModel + class Wait < OpenAI::Internal::Type::BaseModel attr_accessor type: :wait def initialize: (?type: :wait) -> void @@ -234,7 +234,7 @@ module OpenAI type pending_safety_check = { id: String, code: String, message: String } - class PendingSafetyCheck < OpenAI::BaseModel + class PendingSafetyCheck < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor code: String @@ -249,7 +249,7 @@ module OpenAI type status = :in_progress | :completed | :incomplete module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS: :in_progress COMPLETED: :completed @@ -261,7 +261,7 @@ module OpenAI type type_ = :computer_call module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum COMPUTER_CALL: :computer_call diff --git a/sig/openai/models/responses/response_computer_tool_call_output_item.rbs b/sig/openai/models/responses/response_computer_tool_call_output_item.rbs index 793ed85e..32edd0b4 100644 --- a/sig/openai/models/responses/response_computer_tool_call_output_item.rbs +++ b/sig/openai/models/responses/response_computer_tool_call_output_item.rbs @@ -11,7 +11,7 @@ module OpenAI status: OpenAI::Models::Responses::ResponseComputerToolCallOutputItem::status } - class ResponseComputerToolCallOutputItem < OpenAI::BaseModel + class ResponseComputerToolCallOutputItem < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor call_id: String @@ -46,7 +46,7 @@ module OpenAI type acknowledged_safety_check = { id: String, code: String, message: String } - class AcknowledgedSafetyCheck < OpenAI::BaseModel + class AcknowledgedSafetyCheck < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor code: String @@ -61,7 +61,7 @@ module OpenAI type status = :in_progress | :completed | :incomplete module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS: :in_progress COMPLETED: :completed diff --git a/sig/openai/models/responses/response_computer_tool_call_output_screenshot.rbs b/sig/openai/models/responses/response_computer_tool_call_output_screenshot.rbs index 3d8cc77b..eae3e223 100644 --- a/sig/openai/models/responses/response_computer_tool_call_output_screenshot.rbs +++ b/sig/openai/models/responses/response_computer_tool_call_output_screenshot.rbs @@ -4,7 +4,7 @@ module OpenAI type response_computer_tool_call_output_screenshot = { type: :computer_screenshot, file_id: String, image_url: String } - class ResponseComputerToolCallOutputScreenshot < OpenAI::BaseModel + class ResponseComputerToolCallOutputScreenshot < OpenAI::Internal::Type::BaseModel attr_accessor type: :computer_screenshot attr_reader file_id: String? diff --git a/sig/openai/models/responses/response_content.rbs b/sig/openai/models/responses/response_content.rbs index d8469d84..4e48827c 100644 --- a/sig/openai/models/responses/response_content.rbs +++ b/sig/openai/models/responses/response_content.rbs @@ -9,7 +9,7 @@ module OpenAI | OpenAI::Models::Responses::ResponseOutputRefusal module ResponseContent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Responses::ResponseInputText, OpenAI::Models::Responses::ResponseInputImage, OpenAI::Models::Responses::ResponseInputFile, OpenAI::Models::Responses::ResponseOutputText, OpenAI::Models::Responses::ResponseOutputRefusal] end diff --git a/sig/openai/models/responses/response_content_part_added_event.rbs b/sig/openai/models/responses/response_content_part_added_event.rbs index f4ca687c..e4c3e4bf 100644 --- a/sig/openai/models/responses/response_content_part_added_event.rbs +++ b/sig/openai/models/responses/response_content_part_added_event.rbs @@ -10,7 +10,7 @@ module OpenAI type: :"response.content_part.added" } - class ResponseContentPartAddedEvent < OpenAI::BaseModel + class ResponseContentPartAddedEvent < OpenAI::Internal::Type::BaseModel attr_accessor content_index: Integer attr_accessor item_id: String @@ -36,7 +36,7 @@ module OpenAI | OpenAI::Models::Responses::ResponseOutputRefusal module Part - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Responses::ResponseOutputText, OpenAI::Models::Responses::ResponseOutputRefusal] end diff --git a/sig/openai/models/responses/response_content_part_done_event.rbs b/sig/openai/models/responses/response_content_part_done_event.rbs index 4ffbc22b..ff72b2b3 100644 --- a/sig/openai/models/responses/response_content_part_done_event.rbs +++ b/sig/openai/models/responses/response_content_part_done_event.rbs @@ -10,7 +10,7 @@ module OpenAI type: :"response.content_part.done" } - class ResponseContentPartDoneEvent < OpenAI::BaseModel + class ResponseContentPartDoneEvent < OpenAI::Internal::Type::BaseModel attr_accessor content_index: Integer attr_accessor item_id: String @@ -36,7 +36,7 @@ module OpenAI | OpenAI::Models::Responses::ResponseOutputRefusal module Part - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Responses::ResponseOutputText, OpenAI::Models::Responses::ResponseOutputRefusal] end diff --git a/sig/openai/models/responses/response_create_params.rbs b/sig/openai/models/responses/response_create_params.rbs index d4c12563..5e04dd5a 100644 --- a/sig/openai/models/responses/response_create_params.rbs +++ b/sig/openai/models/responses/response_create_params.rbs @@ -23,7 +23,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class ResponseCreateParams < OpenAI::BaseModel + class ResponseCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -101,7 +101,7 @@ module OpenAI type input = String | OpenAI::Models::Responses::response_input module Input - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, OpenAI::Models::Responses::response_input] end @@ -112,7 +112,7 @@ module OpenAI | OpenAI::Models::Responses::ToolChoiceFunction module ToolChoice - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Responses::tool_choice_options, OpenAI::Models::Responses::ToolChoiceTypes, OpenAI::Models::Responses::ToolChoiceFunction] end @@ -120,7 +120,7 @@ module OpenAI type truncation = :auto | :disabled module Truncation - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO: :auto DISABLED: :disabled diff --git a/sig/openai/models/responses/response_created_event.rbs b/sig/openai/models/responses/response_created_event.rbs index bd5a386a..6848ecf7 100644 --- a/sig/openai/models/responses/response_created_event.rbs +++ b/sig/openai/models/responses/response_created_event.rbs @@ -7,7 +7,7 @@ module OpenAI type: :"response.created" } - class ResponseCreatedEvent < OpenAI::BaseModel + class ResponseCreatedEvent < OpenAI::Internal::Type::BaseModel attr_accessor response: OpenAI::Models::Responses::Response attr_accessor type: :"response.created" diff --git a/sig/openai/models/responses/response_delete_params.rbs b/sig/openai/models/responses/response_delete_params.rbs index ff2fd40f..1cedd65a 100644 --- a/sig/openai/models/responses/response_delete_params.rbs +++ b/sig/openai/models/responses/response_delete_params.rbs @@ -4,7 +4,7 @@ module OpenAI type response_delete_params = { } & OpenAI::Internal::Type::request_parameters - class ResponseDeleteParams < OpenAI::BaseModel + class ResponseDeleteParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/responses/response_error.rbs b/sig/openai/models/responses/response_error.rbs index 4187ffad..f9721f01 100644 --- a/sig/openai/models/responses/response_error.rbs +++ b/sig/openai/models/responses/response_error.rbs @@ -7,7 +7,7 @@ module OpenAI message: String } - class ResponseError < OpenAI::BaseModel + class ResponseError < OpenAI::Internal::Type::BaseModel attr_accessor code: OpenAI::Models::Responses::ResponseError::code attr_accessor message: String @@ -40,7 +40,7 @@ module OpenAI | :image_file_not_found module Code - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum SERVER_ERROR: :server_error RATE_LIMIT_EXCEEDED: :rate_limit_exceeded diff --git a/sig/openai/models/responses/response_error_event.rbs b/sig/openai/models/responses/response_error_event.rbs index 2c037b8a..b64588ae 100644 --- a/sig/openai/models/responses/response_error_event.rbs +++ b/sig/openai/models/responses/response_error_event.rbs @@ -4,7 +4,7 @@ module OpenAI type response_error_event = { code: String?, message: String, param: String?, type: :error } - class ResponseErrorEvent < OpenAI::BaseModel + class ResponseErrorEvent < OpenAI::Internal::Type::BaseModel attr_accessor code: String? attr_accessor message: String diff --git a/sig/openai/models/responses/response_failed_event.rbs b/sig/openai/models/responses/response_failed_event.rbs index 5747ae10..f6f26208 100644 --- a/sig/openai/models/responses/response_failed_event.rbs +++ b/sig/openai/models/responses/response_failed_event.rbs @@ -7,7 +7,7 @@ module OpenAI type: :"response.failed" } - class ResponseFailedEvent < OpenAI::BaseModel + class ResponseFailedEvent < OpenAI::Internal::Type::BaseModel attr_accessor response: OpenAI::Models::Responses::Response attr_accessor type: :"response.failed" diff --git a/sig/openai/models/responses/response_file_search_call_completed_event.rbs b/sig/openai/models/responses/response_file_search_call_completed_event.rbs index 069463fe..8249564b 100644 --- a/sig/openai/models/responses/response_file_search_call_completed_event.rbs +++ b/sig/openai/models/responses/response_file_search_call_completed_event.rbs @@ -8,7 +8,7 @@ module OpenAI type: :"response.file_search_call.completed" } - class ResponseFileSearchCallCompletedEvent < OpenAI::BaseModel + class ResponseFileSearchCallCompletedEvent < OpenAI::Internal::Type::BaseModel attr_accessor item_id: String attr_accessor output_index: Integer diff --git a/sig/openai/models/responses/response_file_search_call_in_progress_event.rbs b/sig/openai/models/responses/response_file_search_call_in_progress_event.rbs index e8163f14..5008567b 100644 --- a/sig/openai/models/responses/response_file_search_call_in_progress_event.rbs +++ b/sig/openai/models/responses/response_file_search_call_in_progress_event.rbs @@ -8,7 +8,7 @@ module OpenAI type: :"response.file_search_call.in_progress" } - class ResponseFileSearchCallInProgressEvent < OpenAI::BaseModel + class ResponseFileSearchCallInProgressEvent < OpenAI::Internal::Type::BaseModel attr_accessor item_id: String attr_accessor output_index: Integer diff --git a/sig/openai/models/responses/response_file_search_call_searching_event.rbs b/sig/openai/models/responses/response_file_search_call_searching_event.rbs index 1745ca78..ba475249 100644 --- a/sig/openai/models/responses/response_file_search_call_searching_event.rbs +++ b/sig/openai/models/responses/response_file_search_call_searching_event.rbs @@ -8,7 +8,7 @@ module OpenAI type: :"response.file_search_call.searching" } - class ResponseFileSearchCallSearchingEvent < OpenAI::BaseModel + class ResponseFileSearchCallSearchingEvent < OpenAI::Internal::Type::BaseModel attr_accessor item_id: String attr_accessor output_index: Integer diff --git a/sig/openai/models/responses/response_file_search_tool_call.rbs b/sig/openai/models/responses/response_file_search_tool_call.rbs index c42ff6ab..6e6bfb5c 100644 --- a/sig/openai/models/responses/response_file_search_tool_call.rbs +++ b/sig/openai/models/responses/response_file_search_tool_call.rbs @@ -10,7 +10,7 @@ module OpenAI results: ::Array[OpenAI::Models::Responses::ResponseFileSearchToolCall::Result]? } - class ResponseFileSearchToolCall < OpenAI::BaseModel + class ResponseFileSearchToolCall < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor queries: ::Array[String] @@ -35,7 +35,7 @@ module OpenAI :in_progress | :searching | :completed | :incomplete | :failed module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS: :in_progress SEARCHING: :searching @@ -55,7 +55,7 @@ module OpenAI text: String } - class Result < OpenAI::BaseModel + class Result < OpenAI::Internal::Type::BaseModel attr_accessor attributes: ::Hash[Symbol, OpenAI::Models::Responses::ResponseFileSearchToolCall::Result::attribute]? attr_reader file_id: String? @@ -87,7 +87,7 @@ module OpenAI type attribute = String | Float | bool module Attribute - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, Float, bool] end diff --git a/sig/openai/models/responses/response_format_text_config.rbs b/sig/openai/models/responses/response_format_text_config.rbs index a78a0964..11e70474 100644 --- a/sig/openai/models/responses/response_format_text_config.rbs +++ b/sig/openai/models/responses/response_format_text_config.rbs @@ -7,7 +7,7 @@ module OpenAI | OpenAI::Models::ResponseFormatJSONObject module ResponseFormatTextConfig - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::ResponseFormatText, OpenAI::Models::Responses::ResponseFormatTextJSONSchemaConfig, OpenAI::Models::ResponseFormatJSONObject] end diff --git a/sig/openai/models/responses/response_format_text_json_schema_config.rbs b/sig/openai/models/responses/response_format_text_json_schema_config.rbs index 899ff7bb..f6a08df7 100644 --- a/sig/openai/models/responses/response_format_text_json_schema_config.rbs +++ b/sig/openai/models/responses/response_format_text_json_schema_config.rbs @@ -10,7 +10,7 @@ module OpenAI strict: bool? } - class ResponseFormatTextJSONSchemaConfig < OpenAI::BaseModel + class ResponseFormatTextJSONSchemaConfig < OpenAI::Internal::Type::BaseModel attr_accessor name: String attr_accessor schema: ::Hash[Symbol, top] diff --git a/sig/openai/models/responses/response_function_call_arguments_delta_event.rbs b/sig/openai/models/responses/response_function_call_arguments_delta_event.rbs index afcf88ef..19067caf 100644 --- a/sig/openai/models/responses/response_function_call_arguments_delta_event.rbs +++ b/sig/openai/models/responses/response_function_call_arguments_delta_event.rbs @@ -9,7 +9,7 @@ module OpenAI type: :"response.function_call_arguments.delta" } - class ResponseFunctionCallArgumentsDeltaEvent < OpenAI::BaseModel + class ResponseFunctionCallArgumentsDeltaEvent < OpenAI::Internal::Type::BaseModel attr_accessor delta: String attr_accessor item_id: String diff --git a/sig/openai/models/responses/response_function_call_arguments_done_event.rbs b/sig/openai/models/responses/response_function_call_arguments_done_event.rbs index 5a0dc613..9a6c9073 100644 --- a/sig/openai/models/responses/response_function_call_arguments_done_event.rbs +++ b/sig/openai/models/responses/response_function_call_arguments_done_event.rbs @@ -9,7 +9,7 @@ module OpenAI type: :"response.function_call_arguments.done" } - class ResponseFunctionCallArgumentsDoneEvent < OpenAI::BaseModel + class ResponseFunctionCallArgumentsDoneEvent < OpenAI::Internal::Type::BaseModel attr_accessor arguments: String attr_accessor item_id: String diff --git a/sig/openai/models/responses/response_function_tool_call.rbs b/sig/openai/models/responses/response_function_tool_call.rbs index af876b35..ec454c7d 100644 --- a/sig/openai/models/responses/response_function_tool_call.rbs +++ b/sig/openai/models/responses/response_function_tool_call.rbs @@ -11,7 +11,7 @@ module OpenAI status: OpenAI::Models::Responses::ResponseFunctionToolCall::status } - class ResponseFunctionToolCall < OpenAI::BaseModel + class ResponseFunctionToolCall < OpenAI::Internal::Type::BaseModel attr_accessor arguments: String attr_accessor call_id: String @@ -44,7 +44,7 @@ module OpenAI type status = :in_progress | :completed | :incomplete module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS: :in_progress COMPLETED: :completed diff --git a/sig/openai/models/responses/response_function_tool_call_output_item.rbs b/sig/openai/models/responses/response_function_tool_call_output_item.rbs index 8feafadc..7fd2c526 100644 --- a/sig/openai/models/responses/response_function_tool_call_output_item.rbs +++ b/sig/openai/models/responses/response_function_tool_call_output_item.rbs @@ -10,7 +10,7 @@ module OpenAI status: OpenAI::Models::Responses::ResponseFunctionToolCallOutputItem::status } - class ResponseFunctionToolCallOutputItem < OpenAI::BaseModel + class ResponseFunctionToolCallOutputItem < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor call_id: String @@ -38,7 +38,7 @@ module OpenAI type status = :in_progress | :completed | :incomplete module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS: :in_progress COMPLETED: :completed diff --git a/sig/openai/models/responses/response_function_web_search.rbs b/sig/openai/models/responses/response_function_web_search.rbs index 4b1b18f8..3a2241fe 100644 --- a/sig/openai/models/responses/response_function_web_search.rbs +++ b/sig/openai/models/responses/response_function_web_search.rbs @@ -8,7 +8,7 @@ module OpenAI type: :web_search_call } - class ResponseFunctionWebSearch < OpenAI::BaseModel + class ResponseFunctionWebSearch < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor status: OpenAI::Models::Responses::ResponseFunctionWebSearch::status @@ -26,7 +26,7 @@ module OpenAI type status = :in_progress | :searching | :completed | :failed module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS: :in_progress SEARCHING: :searching diff --git a/sig/openai/models/responses/response_in_progress_event.rbs b/sig/openai/models/responses/response_in_progress_event.rbs index 35845e80..fb29e454 100644 --- a/sig/openai/models/responses/response_in_progress_event.rbs +++ b/sig/openai/models/responses/response_in_progress_event.rbs @@ -7,7 +7,7 @@ module OpenAI type: :"response.in_progress" } - class ResponseInProgressEvent < OpenAI::BaseModel + class ResponseInProgressEvent < OpenAI::Internal::Type::BaseModel attr_accessor response: OpenAI::Models::Responses::Response attr_accessor type: :"response.in_progress" diff --git a/sig/openai/models/responses/response_includable.rbs b/sig/openai/models/responses/response_includable.rbs index 17f0b9d0..0634f434 100644 --- a/sig/openai/models/responses/response_includable.rbs +++ b/sig/openai/models/responses/response_includable.rbs @@ -7,7 +7,7 @@ module OpenAI | :"computer_call_output.output.image_url" module ResponseIncludable - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum FILE_SEARCH_CALL_RESULTS: :"file_search_call.results" MESSAGE_INPUT_IMAGE_IMAGE_URL: :"message.input_image.image_url" diff --git a/sig/openai/models/responses/response_incomplete_event.rbs b/sig/openai/models/responses/response_incomplete_event.rbs index 24d172fe..8e4251ac 100644 --- a/sig/openai/models/responses/response_incomplete_event.rbs +++ b/sig/openai/models/responses/response_incomplete_event.rbs @@ -7,7 +7,7 @@ module OpenAI type: :"response.incomplete" } - class ResponseIncompleteEvent < OpenAI::BaseModel + class ResponseIncompleteEvent < OpenAI::Internal::Type::BaseModel attr_accessor response: OpenAI::Models::Responses::Response attr_accessor type: :"response.incomplete" diff --git a/sig/openai/models/responses/response_input_audio.rbs b/sig/openai/models/responses/response_input_audio.rbs index cebca75a..605fd682 100644 --- a/sig/openai/models/responses/response_input_audio.rbs +++ b/sig/openai/models/responses/response_input_audio.rbs @@ -8,7 +8,7 @@ module OpenAI type: :input_audio } - class ResponseInputAudio < OpenAI::BaseModel + class ResponseInputAudio < OpenAI::Internal::Type::BaseModel attr_accessor data: String attr_accessor format_: OpenAI::Models::Responses::ResponseInputAudio::format_ @@ -26,7 +26,7 @@ module OpenAI type format_ = :mp3 | :wav module Format - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum MP3: :mp3 WAV: :wav diff --git a/sig/openai/models/responses/response_input_content.rbs b/sig/openai/models/responses/response_input_content.rbs index b94f1f20..04d4f7b9 100644 --- a/sig/openai/models/responses/response_input_content.rbs +++ b/sig/openai/models/responses/response_input_content.rbs @@ -7,7 +7,7 @@ module OpenAI | OpenAI::Models::Responses::ResponseInputFile module ResponseInputContent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Responses::ResponseInputText, OpenAI::Models::Responses::ResponseInputImage, OpenAI::Models::Responses::ResponseInputFile] end diff --git a/sig/openai/models/responses/response_input_file.rbs b/sig/openai/models/responses/response_input_file.rbs index 4c336cd6..75f0306c 100644 --- a/sig/openai/models/responses/response_input_file.rbs +++ b/sig/openai/models/responses/response_input_file.rbs @@ -9,7 +9,7 @@ module OpenAI filename: String } - class ResponseInputFile < OpenAI::BaseModel + class ResponseInputFile < OpenAI::Internal::Type::BaseModel attr_accessor type: :input_file attr_reader file_data: String? diff --git a/sig/openai/models/responses/response_input_image.rbs b/sig/openai/models/responses/response_input_image.rbs index 18b3631c..f6e47720 100644 --- a/sig/openai/models/responses/response_input_image.rbs +++ b/sig/openai/models/responses/response_input_image.rbs @@ -9,7 +9,7 @@ module OpenAI image_url: String? } - class ResponseInputImage < OpenAI::BaseModel + class ResponseInputImage < OpenAI::Internal::Type::BaseModel attr_accessor detail: OpenAI::Models::Responses::ResponseInputImage::detail attr_accessor type: :input_image @@ -30,7 +30,7 @@ module OpenAI type detail = :high | :low | :auto module Detail - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum HIGH: :high LOW: :low diff --git a/sig/openai/models/responses/response_input_item.rbs b/sig/openai/models/responses/response_input_item.rbs index ec441979..35ad1d08 100644 --- a/sig/openai/models/responses/response_input_item.rbs +++ b/sig/openai/models/responses/response_input_item.rbs @@ -15,7 +15,7 @@ module OpenAI | OpenAI::Models::Responses::ResponseInputItem::ItemReference module ResponseInputItem - extend OpenAI::Union + extend OpenAI::Internal::Type::Union type message = { @@ -25,7 +25,7 @@ module OpenAI type: OpenAI::Models::Responses::ResponseInputItem::Message::type_ } - class Message < OpenAI::BaseModel + class Message < OpenAI::Internal::Type::BaseModel attr_accessor content: OpenAI::Models::Responses::response_input_message_content_list attr_accessor role: OpenAI::Models::Responses::ResponseInputItem::Message::role @@ -54,7 +54,7 @@ module OpenAI type role = :user | :system | :developer module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum USER: :user SYSTEM: :system @@ -66,7 +66,7 @@ module OpenAI type status = :in_progress | :completed | :incomplete module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS: :in_progress COMPLETED: :completed @@ -78,7 +78,7 @@ module OpenAI type type_ = :message module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum MESSAGE: :message @@ -96,7 +96,7 @@ module OpenAI status: OpenAI::Models::Responses::ResponseInputItem::ComputerCallOutput::status } - class ComputerCallOutput < OpenAI::BaseModel + class ComputerCallOutput < OpenAI::Internal::Type::BaseModel attr_accessor call_id: String attr_accessor output: OpenAI::Models::Responses::ResponseComputerToolCallOutputScreenshot @@ -133,7 +133,7 @@ module OpenAI type acknowledged_safety_check = { id: String, code: String, message: String } - class AcknowledgedSafetyCheck < OpenAI::BaseModel + class AcknowledgedSafetyCheck < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor code: String @@ -148,7 +148,7 @@ module OpenAI type status = :in_progress | :completed | :incomplete module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS: :in_progress COMPLETED: :completed @@ -167,7 +167,7 @@ module OpenAI status: OpenAI::Models::Responses::ResponseInputItem::FunctionCallOutput::status } - class FunctionCallOutput < OpenAI::BaseModel + class FunctionCallOutput < OpenAI::Internal::Type::BaseModel attr_accessor call_id: String attr_accessor output: String @@ -197,7 +197,7 @@ module OpenAI type status = :in_progress | :completed | :incomplete module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS: :in_progress COMPLETED: :completed @@ -209,7 +209,7 @@ module OpenAI type item_reference = { id: String, type: :item_reference } - class ItemReference < OpenAI::BaseModel + class ItemReference < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor type: :item_reference diff --git a/sig/openai/models/responses/response_input_message_item.rbs b/sig/openai/models/responses/response_input_message_item.rbs index e0a6c449..96c1aa3d 100644 --- a/sig/openai/models/responses/response_input_message_item.rbs +++ b/sig/openai/models/responses/response_input_message_item.rbs @@ -10,7 +10,7 @@ module OpenAI type: OpenAI::Models::Responses::ResponseInputMessageItem::type_ } - class ResponseInputMessageItem < OpenAI::BaseModel + class ResponseInputMessageItem < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor content: OpenAI::Models::Responses::response_input_message_content_list @@ -42,7 +42,7 @@ module OpenAI type role = :user | :system | :developer module Role - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum USER: :user SYSTEM: :system @@ -54,7 +54,7 @@ module OpenAI type status = :in_progress | :completed | :incomplete module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS: :in_progress COMPLETED: :completed @@ -66,7 +66,7 @@ module OpenAI type type_ = :message module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum MESSAGE: :message diff --git a/sig/openai/models/responses/response_input_text.rbs b/sig/openai/models/responses/response_input_text.rbs index c41c7537..bb0b03f9 100644 --- a/sig/openai/models/responses/response_input_text.rbs +++ b/sig/openai/models/responses/response_input_text.rbs @@ -3,7 +3,7 @@ module OpenAI module Responses type response_input_text = { text: String, type: :input_text } - class ResponseInputText < OpenAI::BaseModel + class ResponseInputText < OpenAI::Internal::Type::BaseModel attr_accessor text: String attr_accessor type: :input_text diff --git a/sig/openai/models/responses/response_item.rbs b/sig/openai/models/responses/response_item.rbs index 59b6a1f5..a84d2504 100644 --- a/sig/openai/models/responses/response_item.rbs +++ b/sig/openai/models/responses/response_item.rbs @@ -12,7 +12,7 @@ module OpenAI | OpenAI::Models::Responses::ResponseFunctionToolCallOutputItem module ResponseItem - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Responses::ResponseInputMessageItem, OpenAI::Models::Responses::ResponseOutputMessage, OpenAI::Models::Responses::ResponseFileSearchToolCall, OpenAI::Models::Responses::ResponseComputerToolCall, OpenAI::Models::Responses::ResponseComputerToolCallOutputItem, OpenAI::Models::Responses::ResponseFunctionWebSearch, OpenAI::Models::Responses::ResponseFunctionToolCallItem, OpenAI::Models::Responses::ResponseFunctionToolCallOutputItem] end diff --git a/sig/openai/models/responses/response_item_list.rbs b/sig/openai/models/responses/response_item_list.rbs index dbed0410..196112a5 100644 --- a/sig/openai/models/responses/response_item_list.rbs +++ b/sig/openai/models/responses/response_item_list.rbs @@ -12,7 +12,7 @@ module OpenAI object: :list } - class ResponseItemList < OpenAI::BaseModel + class ResponseItemList < OpenAI::Internal::Type::BaseModel attr_accessor data: ::Array[OpenAI::Models::Responses::response_item] attr_accessor first_id: String diff --git a/sig/openai/models/responses/response_output_audio.rbs b/sig/openai/models/responses/response_output_audio.rbs index 1f90e75d..23c68656 100644 --- a/sig/openai/models/responses/response_output_audio.rbs +++ b/sig/openai/models/responses/response_output_audio.rbs @@ -4,7 +4,7 @@ module OpenAI type response_output_audio = { data: String, transcript: String, type: :output_audio } - class ResponseOutputAudio < OpenAI::BaseModel + class ResponseOutputAudio < OpenAI::Internal::Type::BaseModel attr_accessor data: String attr_accessor transcript: String diff --git a/sig/openai/models/responses/response_output_item.rbs b/sig/openai/models/responses/response_output_item.rbs index fb0c190a..f3a5b7b5 100644 --- a/sig/openai/models/responses/response_output_item.rbs +++ b/sig/openai/models/responses/response_output_item.rbs @@ -10,7 +10,7 @@ module OpenAI | OpenAI::Models::Responses::ResponseReasoningItem module ResponseOutputItem - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Responses::ResponseOutputMessage, OpenAI::Models::Responses::ResponseFileSearchToolCall, OpenAI::Models::Responses::ResponseFunctionToolCall, OpenAI::Models::Responses::ResponseFunctionWebSearch, OpenAI::Models::Responses::ResponseComputerToolCall, OpenAI::Models::Responses::ResponseReasoningItem] end diff --git a/sig/openai/models/responses/response_output_item_added_event.rbs b/sig/openai/models/responses/response_output_item_added_event.rbs index e59d12f3..bdbcfb2e 100644 --- a/sig/openai/models/responses/response_output_item_added_event.rbs +++ b/sig/openai/models/responses/response_output_item_added_event.rbs @@ -8,7 +8,7 @@ module OpenAI type: :"response.output_item.added" } - class ResponseOutputItemAddedEvent < OpenAI::BaseModel + class ResponseOutputItemAddedEvent < OpenAI::Internal::Type::BaseModel attr_accessor item: OpenAI::Models::Responses::response_output_item attr_accessor output_index: Integer diff --git a/sig/openai/models/responses/response_output_item_done_event.rbs b/sig/openai/models/responses/response_output_item_done_event.rbs index c4df8604..076a3f09 100644 --- a/sig/openai/models/responses/response_output_item_done_event.rbs +++ b/sig/openai/models/responses/response_output_item_done_event.rbs @@ -8,7 +8,7 @@ module OpenAI type: :"response.output_item.done" } - class ResponseOutputItemDoneEvent < OpenAI::BaseModel + class ResponseOutputItemDoneEvent < OpenAI::Internal::Type::BaseModel attr_accessor item: OpenAI::Models::Responses::response_output_item attr_accessor output_index: Integer diff --git a/sig/openai/models/responses/response_output_message.rbs b/sig/openai/models/responses/response_output_message.rbs index dd1c2a28..8662cc68 100644 --- a/sig/openai/models/responses/response_output_message.rbs +++ b/sig/openai/models/responses/response_output_message.rbs @@ -10,7 +10,7 @@ module OpenAI type: :message } - class ResponseOutputMessage < OpenAI::BaseModel + class ResponseOutputMessage < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor content: ::Array[OpenAI::Models::Responses::ResponseOutputMessage::content] @@ -36,7 +36,7 @@ module OpenAI | OpenAI::Models::Responses::ResponseOutputRefusal module Content - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Responses::ResponseOutputText, OpenAI::Models::Responses::ResponseOutputRefusal] end @@ -44,7 +44,7 @@ module OpenAI type status = :in_progress | :completed | :incomplete module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS: :in_progress COMPLETED: :completed diff --git a/sig/openai/models/responses/response_output_refusal.rbs b/sig/openai/models/responses/response_output_refusal.rbs index f1743e42..69e08eed 100644 --- a/sig/openai/models/responses/response_output_refusal.rbs +++ b/sig/openai/models/responses/response_output_refusal.rbs @@ -3,7 +3,7 @@ module OpenAI module Responses type response_output_refusal = { refusal: String, type: :refusal } - class ResponseOutputRefusal < OpenAI::BaseModel + class ResponseOutputRefusal < OpenAI::Internal::Type::BaseModel attr_accessor refusal: String attr_accessor type: :refusal diff --git a/sig/openai/models/responses/response_output_text.rbs b/sig/openai/models/responses/response_output_text.rbs index 51f31ba9..e3814bc4 100644 --- a/sig/openai/models/responses/response_output_text.rbs +++ b/sig/openai/models/responses/response_output_text.rbs @@ -8,7 +8,7 @@ module OpenAI type: :output_text } - class ResponseOutputText < OpenAI::BaseModel + class ResponseOutputText < OpenAI::Internal::Type::BaseModel attr_accessor annotations: ::Array[OpenAI::Models::Responses::ResponseOutputText::annotation] attr_accessor text: String @@ -29,12 +29,12 @@ module OpenAI | OpenAI::Models::Responses::ResponseOutputText::Annotation::FilePath module Annotation - extend OpenAI::Union + extend OpenAI::Internal::Type::Union type file_citation = { file_id: String, index: Integer, type: :file_citation } - class FileCitation < OpenAI::BaseModel + class FileCitation < OpenAI::Internal::Type::BaseModel attr_accessor file_id: String attr_accessor index: Integer @@ -59,7 +59,7 @@ module OpenAI url: String } - class URLCitation < OpenAI::BaseModel + class URLCitation < OpenAI::Internal::Type::BaseModel attr_accessor end_index: Integer attr_accessor start_index: Integer @@ -83,7 +83,7 @@ module OpenAI type file_path = { file_id: String, index: Integer, type: :file_path } - class FilePath < OpenAI::BaseModel + class FilePath < OpenAI::Internal::Type::BaseModel attr_accessor file_id: String attr_accessor index: Integer diff --git a/sig/openai/models/responses/response_reasoning_item.rbs b/sig/openai/models/responses/response_reasoning_item.rbs index 9c17d2fd..e83fbe06 100644 --- a/sig/openai/models/responses/response_reasoning_item.rbs +++ b/sig/openai/models/responses/response_reasoning_item.rbs @@ -9,7 +9,7 @@ module OpenAI status: OpenAI::Models::Responses::ResponseReasoningItem::status } - class ResponseReasoningItem < OpenAI::BaseModel + class ResponseReasoningItem < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor summary: ::Array[OpenAI::Models::Responses::ResponseReasoningItem::Summary] @@ -33,7 +33,7 @@ module OpenAI type summary = { text: String, type: :summary_text } - class Summary < OpenAI::BaseModel + class Summary < OpenAI::Internal::Type::BaseModel attr_accessor text: String attr_accessor type: :summary_text @@ -46,7 +46,7 @@ module OpenAI type status = :in_progress | :completed | :incomplete module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS: :in_progress COMPLETED: :completed diff --git a/sig/openai/models/responses/response_refusal_delta_event.rbs b/sig/openai/models/responses/response_refusal_delta_event.rbs index 6d26e2d3..0ab5e996 100644 --- a/sig/openai/models/responses/response_refusal_delta_event.rbs +++ b/sig/openai/models/responses/response_refusal_delta_event.rbs @@ -10,7 +10,7 @@ module OpenAI type: :"response.refusal.delta" } - class ResponseRefusalDeltaEvent < OpenAI::BaseModel + class ResponseRefusalDeltaEvent < OpenAI::Internal::Type::BaseModel attr_accessor content_index: Integer attr_accessor delta: String diff --git a/sig/openai/models/responses/response_refusal_done_event.rbs b/sig/openai/models/responses/response_refusal_done_event.rbs index 693e0216..761e42a0 100644 --- a/sig/openai/models/responses/response_refusal_done_event.rbs +++ b/sig/openai/models/responses/response_refusal_done_event.rbs @@ -10,7 +10,7 @@ module OpenAI type: :"response.refusal.done" } - class ResponseRefusalDoneEvent < OpenAI::BaseModel + class ResponseRefusalDoneEvent < OpenAI::Internal::Type::BaseModel attr_accessor content_index: Integer attr_accessor item_id: String diff --git a/sig/openai/models/responses/response_retrieve_params.rbs b/sig/openai/models/responses/response_retrieve_params.rbs index d91ca0ed..9a99b74c 100644 --- a/sig/openai/models/responses/response_retrieve_params.rbs +++ b/sig/openai/models/responses/response_retrieve_params.rbs @@ -5,7 +5,7 @@ module OpenAI { include: ::Array[OpenAI::Models::Responses::response_includable] } & OpenAI::Internal::Type::request_parameters - class ResponseRetrieveParams < OpenAI::BaseModel + class ResponseRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/responses/response_status.rbs b/sig/openai/models/responses/response_status.rbs index cc7c73f5..d78fb3a2 100644 --- a/sig/openai/models/responses/response_status.rbs +++ b/sig/openai/models/responses/response_status.rbs @@ -4,7 +4,7 @@ module OpenAI type response_status = :completed | :failed | :in_progress | :incomplete module ResponseStatus - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum COMPLETED: :completed FAILED: :failed diff --git a/sig/openai/models/responses/response_stream_event.rbs b/sig/openai/models/responses/response_stream_event.rbs index a73d12f4..18708efb 100644 --- a/sig/openai/models/responses/response_stream_event.rbs +++ b/sig/openai/models/responses/response_stream_event.rbs @@ -36,7 +36,7 @@ module OpenAI | OpenAI::Models::Responses::ResponseWebSearchCallSearchingEvent module ResponseStreamEvent - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Responses::ResponseAudioDeltaEvent, OpenAI::Models::Responses::ResponseAudioDoneEvent, OpenAI::Models::Responses::ResponseAudioTranscriptDeltaEvent, OpenAI::Models::Responses::ResponseAudioTranscriptDoneEvent, OpenAI::Models::Responses::ResponseCodeInterpreterCallCodeDeltaEvent, OpenAI::Models::Responses::ResponseCodeInterpreterCallCodeDoneEvent, OpenAI::Models::Responses::ResponseCodeInterpreterCallCompletedEvent, OpenAI::Models::Responses::ResponseCodeInterpreterCallInProgressEvent, OpenAI::Models::Responses::ResponseCodeInterpreterCallInterpretingEvent, OpenAI::Models::Responses::ResponseCompletedEvent, OpenAI::Models::Responses::ResponseContentPartAddedEvent, OpenAI::Models::Responses::ResponseContentPartDoneEvent, OpenAI::Models::Responses::ResponseCreatedEvent, OpenAI::Models::Responses::ResponseErrorEvent, OpenAI::Models::Responses::ResponseFileSearchCallCompletedEvent, OpenAI::Models::Responses::ResponseFileSearchCallInProgressEvent, OpenAI::Models::Responses::ResponseFileSearchCallSearchingEvent, OpenAI::Models::Responses::ResponseFunctionCallArgumentsDeltaEvent, OpenAI::Models::Responses::ResponseFunctionCallArgumentsDoneEvent, OpenAI::Models::Responses::ResponseInProgressEvent, OpenAI::Models::Responses::ResponseFailedEvent, OpenAI::Models::Responses::ResponseIncompleteEvent, OpenAI::Models::Responses::ResponseOutputItemAddedEvent, OpenAI::Models::Responses::ResponseOutputItemDoneEvent, OpenAI::Models::Responses::ResponseRefusalDeltaEvent, OpenAI::Models::Responses::ResponseRefusalDoneEvent, OpenAI::Models::Responses::ResponseTextAnnotationDeltaEvent, OpenAI::Models::Responses::ResponseTextDeltaEvent, OpenAI::Models::Responses::ResponseTextDoneEvent, OpenAI::Models::Responses::ResponseWebSearchCallCompletedEvent, OpenAI::Models::Responses::ResponseWebSearchCallInProgressEvent, OpenAI::Models::Responses::ResponseWebSearchCallSearchingEvent] end diff --git a/sig/openai/models/responses/response_text_annotation_delta_event.rbs b/sig/openai/models/responses/response_text_annotation_delta_event.rbs index 71096d81..54d32aa3 100644 --- a/sig/openai/models/responses/response_text_annotation_delta_event.rbs +++ b/sig/openai/models/responses/response_text_annotation_delta_event.rbs @@ -11,7 +11,7 @@ module OpenAI type: :"response.output_text.annotation.added" } - class ResponseTextAnnotationDeltaEvent < OpenAI::BaseModel + class ResponseTextAnnotationDeltaEvent < OpenAI::Internal::Type::BaseModel attr_accessor annotation: OpenAI::Models::Responses::ResponseTextAnnotationDeltaEvent::annotation attr_accessor annotation_index: Integer @@ -41,12 +41,12 @@ module OpenAI | OpenAI::Models::Responses::ResponseTextAnnotationDeltaEvent::Annotation::FilePath module Annotation - extend OpenAI::Union + extend OpenAI::Internal::Type::Union type file_citation = { file_id: String, index: Integer, type: :file_citation } - class FileCitation < OpenAI::BaseModel + class FileCitation < OpenAI::Internal::Type::BaseModel attr_accessor file_id: String attr_accessor index: Integer @@ -71,7 +71,7 @@ module OpenAI url: String } - class URLCitation < OpenAI::BaseModel + class URLCitation < OpenAI::Internal::Type::BaseModel attr_accessor end_index: Integer attr_accessor start_index: Integer @@ -95,7 +95,7 @@ module OpenAI type file_path = { file_id: String, index: Integer, type: :file_path } - class FilePath < OpenAI::BaseModel + class FilePath < OpenAI::Internal::Type::BaseModel attr_accessor file_id: String attr_accessor index: Integer diff --git a/sig/openai/models/responses/response_text_config.rbs b/sig/openai/models/responses/response_text_config.rbs index 2429e169..30ff82f7 100644 --- a/sig/openai/models/responses/response_text_config.rbs +++ b/sig/openai/models/responses/response_text_config.rbs @@ -4,7 +4,7 @@ module OpenAI type response_text_config = { format_: OpenAI::Models::Responses::response_format_text_config } - class ResponseTextConfig < OpenAI::BaseModel + class ResponseTextConfig < OpenAI::Internal::Type::BaseModel attr_reader format_: OpenAI::Models::Responses::response_format_text_config? def format_=: ( diff --git a/sig/openai/models/responses/response_text_delta_event.rbs b/sig/openai/models/responses/response_text_delta_event.rbs index 45cfac57..2c36a1bf 100644 --- a/sig/openai/models/responses/response_text_delta_event.rbs +++ b/sig/openai/models/responses/response_text_delta_event.rbs @@ -10,7 +10,7 @@ module OpenAI type: :"response.output_text.delta" } - class ResponseTextDeltaEvent < OpenAI::BaseModel + class ResponseTextDeltaEvent < OpenAI::Internal::Type::BaseModel attr_accessor content_index: Integer attr_accessor delta: String diff --git a/sig/openai/models/responses/response_text_done_event.rbs b/sig/openai/models/responses/response_text_done_event.rbs index bc4c94d4..86dc66db 100644 --- a/sig/openai/models/responses/response_text_done_event.rbs +++ b/sig/openai/models/responses/response_text_done_event.rbs @@ -10,7 +10,7 @@ module OpenAI type: :"response.output_text.done" } - class ResponseTextDoneEvent < OpenAI::BaseModel + class ResponseTextDoneEvent < OpenAI::Internal::Type::BaseModel attr_accessor content_index: Integer attr_accessor item_id: String diff --git a/sig/openai/models/responses/response_usage.rbs b/sig/openai/models/responses/response_usage.rbs index 8f4d0dac..28b094d0 100644 --- a/sig/openai/models/responses/response_usage.rbs +++ b/sig/openai/models/responses/response_usage.rbs @@ -10,7 +10,7 @@ module OpenAI total_tokens: Integer } - class ResponseUsage < OpenAI::BaseModel + class ResponseUsage < OpenAI::Internal::Type::BaseModel attr_accessor input_tokens: Integer attr_accessor input_tokens_details: OpenAI::Models::Responses::ResponseUsage::InputTokensDetails @@ -33,7 +33,7 @@ module OpenAI type input_tokens_details = { cached_tokens: Integer } - class InputTokensDetails < OpenAI::BaseModel + class InputTokensDetails < OpenAI::Internal::Type::BaseModel attr_accessor cached_tokens: Integer def initialize: (cached_tokens: Integer) -> void @@ -43,7 +43,7 @@ module OpenAI type output_tokens_details = { reasoning_tokens: Integer } - class OutputTokensDetails < OpenAI::BaseModel + class OutputTokensDetails < OpenAI::Internal::Type::BaseModel attr_accessor reasoning_tokens: Integer def initialize: (reasoning_tokens: Integer) -> void diff --git a/sig/openai/models/responses/response_web_search_call_completed_event.rbs b/sig/openai/models/responses/response_web_search_call_completed_event.rbs index bc8d0e77..d5772178 100644 --- a/sig/openai/models/responses/response_web_search_call_completed_event.rbs +++ b/sig/openai/models/responses/response_web_search_call_completed_event.rbs @@ -8,7 +8,7 @@ module OpenAI type: :"response.web_search_call.completed" } - class ResponseWebSearchCallCompletedEvent < OpenAI::BaseModel + class ResponseWebSearchCallCompletedEvent < OpenAI::Internal::Type::BaseModel attr_accessor item_id: String attr_accessor output_index: Integer diff --git a/sig/openai/models/responses/response_web_search_call_in_progress_event.rbs b/sig/openai/models/responses/response_web_search_call_in_progress_event.rbs index 34cdc0c4..56dfa277 100644 --- a/sig/openai/models/responses/response_web_search_call_in_progress_event.rbs +++ b/sig/openai/models/responses/response_web_search_call_in_progress_event.rbs @@ -8,7 +8,7 @@ module OpenAI type: :"response.web_search_call.in_progress" } - class ResponseWebSearchCallInProgressEvent < OpenAI::BaseModel + class ResponseWebSearchCallInProgressEvent < OpenAI::Internal::Type::BaseModel attr_accessor item_id: String attr_accessor output_index: Integer diff --git a/sig/openai/models/responses/response_web_search_call_searching_event.rbs b/sig/openai/models/responses/response_web_search_call_searching_event.rbs index f03cf1a9..3a70544f 100644 --- a/sig/openai/models/responses/response_web_search_call_searching_event.rbs +++ b/sig/openai/models/responses/response_web_search_call_searching_event.rbs @@ -8,7 +8,7 @@ module OpenAI type: :"response.web_search_call.searching" } - class ResponseWebSearchCallSearchingEvent < OpenAI::BaseModel + class ResponseWebSearchCallSearchingEvent < OpenAI::Internal::Type::BaseModel attr_accessor item_id: String attr_accessor output_index: Integer diff --git a/sig/openai/models/responses/tool.rbs b/sig/openai/models/responses/tool.rbs index c3b27fe5..0d3a572d 100644 --- a/sig/openai/models/responses/tool.rbs +++ b/sig/openai/models/responses/tool.rbs @@ -8,7 +8,7 @@ module OpenAI | OpenAI::Models::Responses::WebSearchTool module Tool - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::Responses::FileSearchTool, OpenAI::Models::Responses::FunctionTool, OpenAI::Models::Responses::ComputerTool, OpenAI::Models::Responses::WebSearchTool] end diff --git a/sig/openai/models/responses/tool_choice_function.rbs b/sig/openai/models/responses/tool_choice_function.rbs index bf83f1a7..a1881b61 100644 --- a/sig/openai/models/responses/tool_choice_function.rbs +++ b/sig/openai/models/responses/tool_choice_function.rbs @@ -3,7 +3,7 @@ module OpenAI module Responses type tool_choice_function = { name: String, type: :function } - class ToolChoiceFunction < OpenAI::BaseModel + class ToolChoiceFunction < OpenAI::Internal::Type::BaseModel attr_accessor name: String attr_accessor type: :function diff --git a/sig/openai/models/responses/tool_choice_options.rbs b/sig/openai/models/responses/tool_choice_options.rbs index 44974540..f07ff227 100644 --- a/sig/openai/models/responses/tool_choice_options.rbs +++ b/sig/openai/models/responses/tool_choice_options.rbs @@ -4,7 +4,7 @@ module OpenAI type tool_choice_options = :none | :auto | :required module ToolChoiceOptions - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum NONE: :none AUTO: :auto diff --git a/sig/openai/models/responses/tool_choice_types.rbs b/sig/openai/models/responses/tool_choice_types.rbs index b0562635..588a58ac 100644 --- a/sig/openai/models/responses/tool_choice_types.rbs +++ b/sig/openai/models/responses/tool_choice_types.rbs @@ -4,7 +4,7 @@ module OpenAI type tool_choice_types = { type: OpenAI::Models::Responses::ToolChoiceTypes::type_ } - class ToolChoiceTypes < OpenAI::BaseModel + class ToolChoiceTypes < OpenAI::Internal::Type::BaseModel attr_accessor type: OpenAI::Models::Responses::ToolChoiceTypes::type_ def initialize: ( @@ -20,7 +20,7 @@ module OpenAI | :web_search_preview_2025_03_11 module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum FILE_SEARCH: :file_search WEB_SEARCH_PREVIEW: :web_search_preview diff --git a/sig/openai/models/responses/web_search_tool.rbs b/sig/openai/models/responses/web_search_tool.rbs index cb6606be..92f25bd0 100644 --- a/sig/openai/models/responses/web_search_tool.rbs +++ b/sig/openai/models/responses/web_search_tool.rbs @@ -8,7 +8,7 @@ module OpenAI user_location: OpenAI::Models::Responses::WebSearchTool::UserLocation? } - class WebSearchTool < OpenAI::BaseModel + class WebSearchTool < OpenAI::Internal::Type::BaseModel attr_accessor type: OpenAI::Models::Responses::WebSearchTool::type_ attr_reader search_context_size: OpenAI::Models::Responses::WebSearchTool::search_context_size? @@ -30,7 +30,7 @@ module OpenAI type type_ = :web_search_preview | :web_search_preview_2025_03_11 module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum WEB_SEARCH_PREVIEW: :web_search_preview WEB_SEARCH_PREVIEW_2025_03_11: :web_search_preview_2025_03_11 @@ -41,7 +41,7 @@ module OpenAI type search_context_size = :low | :medium | :high module SearchContextSize - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum LOW: :low MEDIUM: :medium @@ -59,7 +59,7 @@ module OpenAI timezone: String } - class UserLocation < OpenAI::BaseModel + class UserLocation < OpenAI::Internal::Type::BaseModel attr_accessor type: :approximate attr_reader city: String? diff --git a/sig/openai/models/responses_model.rbs b/sig/openai/models/responses_model.rbs index 5e0a2660..2cb153e7 100644 --- a/sig/openai/models/responses_model.rbs +++ b/sig/openai/models/responses_model.rbs @@ -9,7 +9,7 @@ module OpenAI | :"computer-use-preview-2025-03-11" module ResponsesModel - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, OpenAI::Models::chat_model, (:"o1-pro" | :"o1-pro-2025-03-19" diff --git a/sig/openai/models/static_file_chunking_strategy.rbs b/sig/openai/models/static_file_chunking_strategy.rbs index e77bd4ac..7c00abb0 100644 --- a/sig/openai/models/static_file_chunking_strategy.rbs +++ b/sig/openai/models/static_file_chunking_strategy.rbs @@ -3,7 +3,7 @@ module OpenAI type static_file_chunking_strategy = { chunk_overlap_tokens: Integer, max_chunk_size_tokens: Integer } - class StaticFileChunkingStrategy < OpenAI::BaseModel + class StaticFileChunkingStrategy < OpenAI::Internal::Type::BaseModel attr_accessor chunk_overlap_tokens: Integer attr_accessor max_chunk_size_tokens: Integer diff --git a/sig/openai/models/static_file_chunking_strategy_object.rbs b/sig/openai/models/static_file_chunking_strategy_object.rbs index 62e506a3..1122e914 100644 --- a/sig/openai/models/static_file_chunking_strategy_object.rbs +++ b/sig/openai/models/static_file_chunking_strategy_object.rbs @@ -3,7 +3,7 @@ module OpenAI type static_file_chunking_strategy_object = { static: OpenAI::Models::StaticFileChunkingStrategy, type: :static } - class StaticFileChunkingStrategyObject < OpenAI::BaseModel + class StaticFileChunkingStrategyObject < OpenAI::Internal::Type::BaseModel attr_accessor static: OpenAI::Models::StaticFileChunkingStrategy attr_accessor type: :static diff --git a/sig/openai/models/static_file_chunking_strategy_object_param.rbs b/sig/openai/models/static_file_chunking_strategy_object_param.rbs index a86f96ab..639a7ea1 100644 --- a/sig/openai/models/static_file_chunking_strategy_object_param.rbs +++ b/sig/openai/models/static_file_chunking_strategy_object_param.rbs @@ -3,7 +3,7 @@ module OpenAI type static_file_chunking_strategy_object_param = { static: OpenAI::Models::StaticFileChunkingStrategy, type: :static } - class StaticFileChunkingStrategyObjectParam < OpenAI::BaseModel + class StaticFileChunkingStrategyObjectParam < OpenAI::Internal::Type::BaseModel attr_accessor static: OpenAI::Models::StaticFileChunkingStrategy attr_accessor type: :static diff --git a/sig/openai/models/upload.rbs b/sig/openai/models/upload.rbs index ff4fec3b..0a50164d 100644 --- a/sig/openai/models/upload.rbs +++ b/sig/openai/models/upload.rbs @@ -13,7 +13,7 @@ module OpenAI file: OpenAI::Models::FileObject? } - class Upload < OpenAI::BaseModel + class Upload < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor bytes: Integer @@ -49,7 +49,7 @@ module OpenAI type status = :pending | :completed | :cancelled | :expired module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum PENDING: :pending COMPLETED: :completed diff --git a/sig/openai/models/upload_cancel_params.rbs b/sig/openai/models/upload_cancel_params.rbs index f7d03f06..669ffc83 100644 --- a/sig/openai/models/upload_cancel_params.rbs +++ b/sig/openai/models/upload_cancel_params.rbs @@ -3,7 +3,7 @@ module OpenAI type upload_cancel_params = { } & OpenAI::Internal::Type::request_parameters - class UploadCancelParams < OpenAI::BaseModel + class UploadCancelParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/upload_complete_params.rbs b/sig/openai/models/upload_complete_params.rbs index 50d22376..5cfff932 100644 --- a/sig/openai/models/upload_complete_params.rbs +++ b/sig/openai/models/upload_complete_params.rbs @@ -4,7 +4,7 @@ module OpenAI { part_ids: ::Array[String], :md5 => String } & OpenAI::Internal::Type::request_parameters - class UploadCompleteParams < OpenAI::BaseModel + class UploadCompleteParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/upload_create_params.rbs b/sig/openai/models/upload_create_params.rbs index 94f2cc72..9b68eb00 100644 --- a/sig/openai/models/upload_create_params.rbs +++ b/sig/openai/models/upload_create_params.rbs @@ -9,7 +9,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class UploadCreateParams < OpenAI::BaseModel + class UploadCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/uploads/part_create_params.rbs b/sig/openai/models/uploads/part_create_params.rbs index a36ae697..d6014118 100644 --- a/sig/openai/models/uploads/part_create_params.rbs +++ b/sig/openai/models/uploads/part_create_params.rbs @@ -4,7 +4,7 @@ module OpenAI type part_create_params = { data: (IO | StringIO) } & OpenAI::Internal::Type::request_parameters - class PartCreateParams < OpenAI::BaseModel + class PartCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/uploads/upload_part.rbs b/sig/openai/models/uploads/upload_part.rbs index 2a1f9e38..93530d14 100644 --- a/sig/openai/models/uploads/upload_part.rbs +++ b/sig/openai/models/uploads/upload_part.rbs @@ -11,7 +11,7 @@ module OpenAI upload_id: String } - class UploadPart < OpenAI::BaseModel + class UploadPart < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor created_at: Integer diff --git a/sig/openai/models/vector_store.rbs b/sig/openai/models/vector_store.rbs index 02c2f02f..a29d28eb 100644 --- a/sig/openai/models/vector_store.rbs +++ b/sig/openai/models/vector_store.rbs @@ -15,7 +15,7 @@ module OpenAI expires_at: Integer? } - class VectorStore < OpenAI::BaseModel + class VectorStore < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor created_at: Integer @@ -67,7 +67,7 @@ module OpenAI total: Integer } - class FileCounts < OpenAI::BaseModel + class FileCounts < OpenAI::Internal::Type::BaseModel attr_accessor cancelled: Integer attr_accessor completed: Integer @@ -92,7 +92,7 @@ module OpenAI type status = :expired | :in_progress | :completed module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum EXPIRED: :expired IN_PROGRESS: :in_progress @@ -103,7 +103,7 @@ module OpenAI type expires_after = { anchor: :last_active_at, days: Integer } - class ExpiresAfter < OpenAI::BaseModel + class ExpiresAfter < OpenAI::Internal::Type::BaseModel attr_accessor anchor: :last_active_at attr_accessor days: Integer diff --git a/sig/openai/models/vector_store_create_params.rbs b/sig/openai/models/vector_store_create_params.rbs index 3a387fa9..accb9698 100644 --- a/sig/openai/models/vector_store_create_params.rbs +++ b/sig/openai/models/vector_store_create_params.rbs @@ -10,7 +10,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class VectorStoreCreateParams < OpenAI::BaseModel + class VectorStoreCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -49,7 +49,7 @@ module OpenAI type expires_after = { anchor: :last_active_at, days: Integer } - class ExpiresAfter < OpenAI::BaseModel + class ExpiresAfter < OpenAI::Internal::Type::BaseModel attr_accessor anchor: :last_active_at attr_accessor days: Integer diff --git a/sig/openai/models/vector_store_delete_params.rbs b/sig/openai/models/vector_store_delete_params.rbs index b2f2436c..1dfb12b4 100644 --- a/sig/openai/models/vector_store_delete_params.rbs +++ b/sig/openai/models/vector_store_delete_params.rbs @@ -3,7 +3,7 @@ module OpenAI type vector_store_delete_params = { } & OpenAI::Internal::Type::request_parameters - class VectorStoreDeleteParams < OpenAI::BaseModel + class VectorStoreDeleteParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/vector_store_deleted.rbs b/sig/openai/models/vector_store_deleted.rbs index 3403e058..1c8ca3ae 100644 --- a/sig/openai/models/vector_store_deleted.rbs +++ b/sig/openai/models/vector_store_deleted.rbs @@ -3,7 +3,7 @@ module OpenAI type vector_store_deleted = { id: String, deleted: bool, object: :"vector_store.deleted" } - class VectorStoreDeleted < OpenAI::BaseModel + class VectorStoreDeleted < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor deleted: bool diff --git a/sig/openai/models/vector_store_list_params.rbs b/sig/openai/models/vector_store_list_params.rbs index 45da51d8..5c2aa3f9 100644 --- a/sig/openai/models/vector_store_list_params.rbs +++ b/sig/openai/models/vector_store_list_params.rbs @@ -9,7 +9,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class VectorStoreListParams < OpenAI::BaseModel + class VectorStoreListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -44,7 +44,7 @@ module OpenAI type order = :asc | :desc module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASC: :asc DESC: :desc diff --git a/sig/openai/models/vector_store_retrieve_params.rbs b/sig/openai/models/vector_store_retrieve_params.rbs index 7f875a4d..51b51332 100644 --- a/sig/openai/models/vector_store_retrieve_params.rbs +++ b/sig/openai/models/vector_store_retrieve_params.rbs @@ -3,7 +3,7 @@ module OpenAI type vector_store_retrieve_params = { } & OpenAI::Internal::Type::request_parameters - class VectorStoreRetrieveParams < OpenAI::BaseModel + class VectorStoreRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/vector_store_search_params.rbs b/sig/openai/models/vector_store_search_params.rbs index 10fb00d0..823cd7bc 100644 --- a/sig/openai/models/vector_store_search_params.rbs +++ b/sig/openai/models/vector_store_search_params.rbs @@ -10,7 +10,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class VectorStoreSearchParams < OpenAI::BaseModel + class VectorStoreSearchParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -50,7 +50,7 @@ module OpenAI type query = String | ::Array[String] module Query - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, ::Array[String]] @@ -63,7 +63,7 @@ module OpenAI OpenAI::Models::ComparisonFilter | OpenAI::Models::CompoundFilter module Filters - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [OpenAI::Models::ComparisonFilter, OpenAI::Models::CompoundFilter] end @@ -74,7 +74,7 @@ module OpenAI score_threshold: Float } - class RankingOptions < OpenAI::BaseModel + class RankingOptions < OpenAI::Internal::Type::BaseModel attr_reader ranker: OpenAI::Models::VectorStoreSearchParams::RankingOptions::ranker? def ranker=: ( @@ -95,7 +95,7 @@ module OpenAI type ranker = :auto | :"default-2024-11-15" module Ranker - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum AUTO: :auto DEFAULT_2024_11_15: :"default-2024-11-15" diff --git a/sig/openai/models/vector_store_search_response.rbs b/sig/openai/models/vector_store_search_response.rbs index a5ba79e3..d7c590bd 100644 --- a/sig/openai/models/vector_store_search_response.rbs +++ b/sig/openai/models/vector_store_search_response.rbs @@ -9,7 +9,7 @@ module OpenAI score: Float } - class VectorStoreSearchResponse < OpenAI::BaseModel + class VectorStoreSearchResponse < OpenAI::Internal::Type::BaseModel attr_accessor attributes: ::Hash[Symbol, OpenAI::Models::VectorStoreSearchResponse::attribute]? attr_accessor content: ::Array[OpenAI::Models::VectorStoreSearchResponse::Content] @@ -33,7 +33,7 @@ module OpenAI type attribute = String | Float | bool module Attribute - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, Float, bool] end @@ -44,7 +44,7 @@ module OpenAI type: OpenAI::Models::VectorStoreSearchResponse::Content::type_ } - class Content < OpenAI::BaseModel + class Content < OpenAI::Internal::Type::BaseModel attr_accessor text: String attr_accessor type: OpenAI::Models::VectorStoreSearchResponse::Content::type_ @@ -59,7 +59,7 @@ module OpenAI type type_ = :text module Type - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TEXT: :text diff --git a/sig/openai/models/vector_store_update_params.rbs b/sig/openai/models/vector_store_update_params.rbs index a25ec5db..f8258d41 100644 --- a/sig/openai/models/vector_store_update_params.rbs +++ b/sig/openai/models/vector_store_update_params.rbs @@ -8,7 +8,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class VectorStoreUpdateParams < OpenAI::BaseModel + class VectorStoreUpdateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -29,7 +29,7 @@ module OpenAI type expires_after = { anchor: :last_active_at, days: Integer } - class ExpiresAfter < OpenAI::BaseModel + class ExpiresAfter < OpenAI::Internal::Type::BaseModel attr_accessor anchor: :last_active_at attr_accessor days: Integer diff --git a/sig/openai/models/vector_stores/file_batch_cancel_params.rbs b/sig/openai/models/vector_stores/file_batch_cancel_params.rbs index eb2088fd..26c9d893 100644 --- a/sig/openai/models/vector_stores/file_batch_cancel_params.rbs +++ b/sig/openai/models/vector_stores/file_batch_cancel_params.rbs @@ -4,7 +4,7 @@ module OpenAI type file_batch_cancel_params = { vector_store_id: String } & OpenAI::Internal::Type::request_parameters - class FileBatchCancelParams < OpenAI::BaseModel + class FileBatchCancelParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/vector_stores/file_batch_create_params.rbs b/sig/openai/models/vector_stores/file_batch_create_params.rbs index e8e0ddca..03b56b5a 100644 --- a/sig/openai/models/vector_stores/file_batch_create_params.rbs +++ b/sig/openai/models/vector_stores/file_batch_create_params.rbs @@ -9,7 +9,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class FileBatchCreateParams < OpenAI::BaseModel + class FileBatchCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -35,7 +35,7 @@ module OpenAI type attribute = String | Float | bool module Attribute - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, Float, bool] end diff --git a/sig/openai/models/vector_stores/file_batch_list_files_params.rbs b/sig/openai/models/vector_stores/file_batch_list_files_params.rbs index e9f2d88b..87868689 100644 --- a/sig/openai/models/vector_stores/file_batch_list_files_params.rbs +++ b/sig/openai/models/vector_stores/file_batch_list_files_params.rbs @@ -12,7 +12,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class FileBatchListFilesParams < OpenAI::BaseModel + class FileBatchListFilesParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -57,7 +57,7 @@ module OpenAI type filter = :in_progress | :completed | :failed | :cancelled module Filter - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS: :in_progress COMPLETED: :completed @@ -70,7 +70,7 @@ module OpenAI type order = :asc | :desc module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASC: :asc DESC: :desc diff --git a/sig/openai/models/vector_stores/file_batch_retrieve_params.rbs b/sig/openai/models/vector_stores/file_batch_retrieve_params.rbs index a9efa644..b43f76b8 100644 --- a/sig/openai/models/vector_stores/file_batch_retrieve_params.rbs +++ b/sig/openai/models/vector_stores/file_batch_retrieve_params.rbs @@ -4,7 +4,7 @@ module OpenAI type file_batch_retrieve_params = { vector_store_id: String } & OpenAI::Internal::Type::request_parameters - class FileBatchRetrieveParams < OpenAI::BaseModel + class FileBatchRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/vector_stores/file_content_params.rbs b/sig/openai/models/vector_stores/file_content_params.rbs index c7f2dfc8..a3f37f30 100644 --- a/sig/openai/models/vector_stores/file_content_params.rbs +++ b/sig/openai/models/vector_stores/file_content_params.rbs @@ -4,7 +4,7 @@ module OpenAI type file_content_params = { vector_store_id: String } & OpenAI::Internal::Type::request_parameters - class FileContentParams < OpenAI::BaseModel + class FileContentParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/vector_stores/file_content_response.rbs b/sig/openai/models/vector_stores/file_content_response.rbs index 11f49799..df2aa45a 100644 --- a/sig/openai/models/vector_stores/file_content_response.rbs +++ b/sig/openai/models/vector_stores/file_content_response.rbs @@ -3,7 +3,7 @@ module OpenAI module VectorStores type file_content_response = { text: String, type: String } - class FileContentResponse < OpenAI::BaseModel + class FileContentResponse < OpenAI::Internal::Type::BaseModel attr_reader text: String? def text=: (String) -> String diff --git a/sig/openai/models/vector_stores/file_create_params.rbs b/sig/openai/models/vector_stores/file_create_params.rbs index 93f6ca13..9a396ee0 100644 --- a/sig/openai/models/vector_stores/file_create_params.rbs +++ b/sig/openai/models/vector_stores/file_create_params.rbs @@ -9,7 +9,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class FileCreateParams < OpenAI::BaseModel + class FileCreateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -35,7 +35,7 @@ module OpenAI type attribute = String | Float | bool module Attribute - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, Float, bool] end diff --git a/sig/openai/models/vector_stores/file_delete_params.rbs b/sig/openai/models/vector_stores/file_delete_params.rbs index 486986b3..a80ed26d 100644 --- a/sig/openai/models/vector_stores/file_delete_params.rbs +++ b/sig/openai/models/vector_stores/file_delete_params.rbs @@ -4,7 +4,7 @@ module OpenAI type file_delete_params = { vector_store_id: String } & OpenAI::Internal::Type::request_parameters - class FileDeleteParams < OpenAI::BaseModel + class FileDeleteParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/vector_stores/file_list_params.rbs b/sig/openai/models/vector_stores/file_list_params.rbs index 6501043b..7b916738 100644 --- a/sig/openai/models/vector_stores/file_list_params.rbs +++ b/sig/openai/models/vector_stores/file_list_params.rbs @@ -11,7 +11,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class FileListParams < OpenAI::BaseModel + class FileListParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -53,7 +53,7 @@ module OpenAI type filter = :in_progress | :completed | :failed | :cancelled module Filter - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS: :in_progress COMPLETED: :completed @@ -66,7 +66,7 @@ module OpenAI type order = :asc | :desc module Order - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ASC: :asc DESC: :desc diff --git a/sig/openai/models/vector_stores/file_retrieve_params.rbs b/sig/openai/models/vector_stores/file_retrieve_params.rbs index f9c4eac7..266f9e3c 100644 --- a/sig/openai/models/vector_stores/file_retrieve_params.rbs +++ b/sig/openai/models/vector_stores/file_retrieve_params.rbs @@ -4,7 +4,7 @@ module OpenAI type file_retrieve_params = { vector_store_id: String } & OpenAI::Internal::Type::request_parameters - class FileRetrieveParams < OpenAI::BaseModel + class FileRetrieveParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters diff --git a/sig/openai/models/vector_stores/file_update_params.rbs b/sig/openai/models/vector_stores/file_update_params.rbs index 3155411a..7b75e4f2 100644 --- a/sig/openai/models/vector_stores/file_update_params.rbs +++ b/sig/openai/models/vector_stores/file_update_params.rbs @@ -8,7 +8,7 @@ module OpenAI } & OpenAI::Internal::Type::request_parameters - class FileUpdateParams < OpenAI::BaseModel + class FileUpdateParams < OpenAI::Internal::Type::BaseModel extend OpenAI::Internal::Type::RequestParameters::Converter include OpenAI::Internal::Type::RequestParameters @@ -27,7 +27,7 @@ module OpenAI type attribute = String | Float | bool module Attribute - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, Float, bool] end diff --git a/sig/openai/models/vector_stores/vector_store_file.rbs b/sig/openai/models/vector_stores/vector_store_file.rbs index 200981e8..7039aa57 100644 --- a/sig/openai/models/vector_stores/vector_store_file.rbs +++ b/sig/openai/models/vector_stores/vector_store_file.rbs @@ -16,7 +16,7 @@ module OpenAI chunking_strategy: OpenAI::Models::file_chunking_strategy } - class VectorStoreFile < OpenAI::BaseModel + class VectorStoreFile < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor created_at: Integer @@ -59,7 +59,7 @@ module OpenAI message: String } - class LastError < OpenAI::BaseModel + class LastError < OpenAI::Internal::Type::BaseModel attr_accessor code: OpenAI::Models::VectorStores::VectorStoreFile::LastError::code attr_accessor message: String @@ -74,7 +74,7 @@ module OpenAI type code = :server_error | :unsupported_file | :invalid_file module Code - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum SERVER_ERROR: :server_error UNSUPPORTED_FILE: :unsupported_file @@ -87,7 +87,7 @@ module OpenAI type status = :in_progress | :completed | :cancelled | :failed module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS: :in_progress COMPLETED: :completed @@ -100,7 +100,7 @@ module OpenAI type attribute = String | Float | bool module Attribute - extend OpenAI::Union + extend OpenAI::Internal::Type::Union def self?.variants: -> [String, Float, bool] end diff --git a/sig/openai/models/vector_stores/vector_store_file_batch.rbs b/sig/openai/models/vector_stores/vector_store_file_batch.rbs index d5ae501f..6c286435 100644 --- a/sig/openai/models/vector_stores/vector_store_file_batch.rbs +++ b/sig/openai/models/vector_stores/vector_store_file_batch.rbs @@ -13,7 +13,7 @@ module OpenAI vector_store_id: String } - class VectorStoreFileBatch < OpenAI::BaseModel + class VectorStoreFileBatch < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor created_at: Integer @@ -46,7 +46,7 @@ module OpenAI total: Integer } - class FileCounts < OpenAI::BaseModel + class FileCounts < OpenAI::Internal::Type::BaseModel attr_accessor cancelled: Integer attr_accessor completed: Integer @@ -71,7 +71,7 @@ module OpenAI type status = :in_progress | :completed | :cancelled | :failed module Status - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum IN_PROGRESS: :in_progress COMPLETED: :completed diff --git a/sig/openai/models/vector_stores/vector_store_file_deleted.rbs b/sig/openai/models/vector_stores/vector_store_file_deleted.rbs index 19257a53..15ae080f 100644 --- a/sig/openai/models/vector_stores/vector_store_file_deleted.rbs +++ b/sig/openai/models/vector_stores/vector_store_file_deleted.rbs @@ -6,7 +6,7 @@ module OpenAI type vector_store_file_deleted = { id: String, deleted: bool, object: :"vector_store.file.deleted" } - class VectorStoreFileDeleted < OpenAI::BaseModel + class VectorStoreFileDeleted < OpenAI::Internal::Type::BaseModel attr_accessor id: String attr_accessor deleted: bool diff --git a/sig/openai/request_options.rbs b/sig/openai/request_options.rbs index 51674ffe..a3ac246f 100644 --- a/sig/openai/request_options.rbs +++ b/sig/openai/request_options.rbs @@ -12,7 +12,7 @@ module OpenAI timeout: Float? } - class RequestOptions < OpenAI::BaseModel + class RequestOptions < OpenAI::Internal::Type::BaseModel def self.validate!: (self | ::Hash[Symbol, top] opts) -> void attr_accessor idempotency_key: String? diff --git a/test/openai/client_test.rb b/test/openai/client_test.rb index 6b35af97..dd6ec3ea 100644 --- a/test/openai/client_test.rb +++ b/test/openai/client_test.rb @@ -55,7 +55,7 @@ def test_client_default_request_default_retry_attempts requester = MockRequester.new(500, {}, {}) openai.requester = requester - assert_raises(OpenAI::InternalServerError) do + assert_raises(OpenAI::Errors::InternalServerError) do openai.chat.completions.create(messages: [{content: "string", role: :developer}], model: :"o3-mini") end @@ -67,7 +67,7 @@ def test_client_given_request_default_retry_attempts requester = MockRequester.new(500, {}, {}) openai.requester = requester - assert_raises(OpenAI::InternalServerError) do + assert_raises(OpenAI::Errors::InternalServerError) do openai.chat.completions.create(messages: [{content: "string", role: :developer}], model: :"o3-mini") end @@ -79,7 +79,7 @@ def test_client_default_request_given_retry_attempts requester = MockRequester.new(500, {}, {}) openai.requester = requester - assert_raises(OpenAI::InternalServerError) do + assert_raises(OpenAI::Errors::InternalServerError) do openai.chat.completions.create( messages: [{content: "string", role: :developer}], model: :"o3-mini", @@ -95,7 +95,7 @@ def test_client_given_request_given_retry_attempts requester = MockRequester.new(500, {}, {}) openai.requester = requester - assert_raises(OpenAI::InternalServerError) do + assert_raises(OpenAI::Errors::InternalServerError) do openai.chat.completions.create( messages: [{content: "string", role: :developer}], model: :"o3-mini", @@ -111,7 +111,7 @@ def test_client_retry_after_seconds requester = MockRequester.new(500, {"retry-after" => "1.3"}, {}) openai.requester = requester - assert_raises(OpenAI::InternalServerError) do + assert_raises(OpenAI::Errors::InternalServerError) do openai.chat.completions.create(messages: [{content: "string", role: :developer}], model: :"o3-mini") end @@ -124,7 +124,7 @@ def test_client_retry_after_date requester = MockRequester.new(500, {"retry-after" => (Time.now + 10).httpdate}, {}) openai.requester = requester - assert_raises(OpenAI::InternalServerError) do + assert_raises(OpenAI::Errors::InternalServerError) do Thread.current.thread_variable_set(:time_now, Time.now) openai.chat.completions.create(messages: [{content: "string", role: :developer}], model: :"o3-mini") Thread.current.thread_variable_set(:time_now, nil) @@ -139,7 +139,7 @@ def test_client_retry_after_ms requester = MockRequester.new(500, {"retry-after-ms" => "1300"}, {}) openai.requester = requester - assert_raises(OpenAI::InternalServerError) do + assert_raises(OpenAI::Errors::InternalServerError) do openai.chat.completions.create(messages: [{content: "string", role: :developer}], model: :"o3-mini") end @@ -152,7 +152,7 @@ def test_retry_count_header requester = MockRequester.new(500, {}, {}) openai.requester = requester - assert_raises(OpenAI::InternalServerError) do + assert_raises(OpenAI::Errors::InternalServerError) do openai.chat.completions.create(messages: [{content: "string", role: :developer}], model: :"o3-mini") end @@ -165,7 +165,7 @@ def test_omit_retry_count_header requester = MockRequester.new(500, {}, {}) openai.requester = requester - assert_raises(OpenAI::InternalServerError) do + assert_raises(OpenAI::Errors::InternalServerError) do openai.chat.completions.create( messages: [{content: "string", role: :developer}], model: :"o3-mini", @@ -182,7 +182,7 @@ def test_overwrite_retry_count_header requester = MockRequester.new(500, {}, {}) openai.requester = requester - assert_raises(OpenAI::InternalServerError) do + assert_raises(OpenAI::Errors::InternalServerError) do openai.chat.completions.create( messages: [{content: "string", role: :developer}], model: :"o3-mini", diff --git a/test/openai/internal/type/base_model_test.rb b/test/openai/internal/type/base_model_test.rb index 25c1322c..0b854098 100644 --- a/test/openai/internal/type/base_model_test.rb +++ b/test/openai/internal/type/base_model_test.rb @@ -3,26 +3,26 @@ require_relative "../../test_helper" class OpenAI::Test::PrimitiveModelTest < Minitest::Test - A = OpenAI::ArrayOf[-> { Integer }] - H = OpenAI::HashOf[-> { Integer }, nil?: true] + A = OpenAI::Internal::Type::ArrayOf[-> { Integer }] + H = OpenAI::Internal::Type::HashOf[-> { Integer }, nil?: true] module E - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum end module U - extend OpenAI::Union + extend OpenAI::Internal::Type::Union end - class B < OpenAI::BaseModel + class B < OpenAI::Internal::Type::BaseModel optional :a, Integer optional :b, B end def test_typing converters = [ - OpenAI::Unknown, - OpenAI::BooleanModel, + OpenAI::Internal::Type::Unknown, + OpenAI::Internal::Type::BooleanModel, A, H, E, @@ -39,11 +39,11 @@ def test_typing def test_coerce cases = { - [OpenAI::Unknown, :a] => [{yes: 1}, :a], + [OpenAI::Internal::Type::Unknown, :a] => [{yes: 1}, :a], [NilClass, :a] => [{maybe: 1}, nil], [NilClass, nil] => [{yes: 1}, nil], - [OpenAI::BooleanModel, true] => [{yes: 1}, true], - [OpenAI::BooleanModel, "true"] => [{no: 1}, "true"], + [OpenAI::Internal::Type::BooleanModel, true] => [{yes: 1}, true], + [OpenAI::Internal::Type::BooleanModel, "true"] => [{no: 1}, "true"], [Integer, 1] => [{yes: 1}, 1], [Integer, 1.0] => [{maybe: 1}, 1], [Integer, "1"] => [{maybe: 1}, 1], @@ -76,7 +76,7 @@ def test_coerce def test_dump cases = { - [OpenAI::Unknown, B.new(a: "one", b: B.new(a: 1.0))] => {a: "one", b: {a: 1}}, + [OpenAI::Internal::Type::Unknown, B.new(a: "one", b: B.new(a: 1.0))] => {a: "one", b: {a: 1}}, [A, B.new(a: "one", b: B.new(a: 1.0))] => {a: "one", b: {a: 1}}, [H, B.new(a: "one", b: B.new(a: 1.0))] => {a: "one", b: {a: 1}}, [E, B.new(a: "one", b: B.new(a: 1.0))] => {a: "one", b: {a: 1}}, @@ -85,8 +85,8 @@ def test_dump [String, B.new(a: "one", b: B.new(a: 1.0))] => {a: "one", b: {a: 1}}, [:b, B.new(a: "one", b: B.new(a: 1.0))] => {a: "one", b: {a: 1}}, [nil, B.new(a: "one", b: B.new(a: 1.0))] => {a: "one", b: {a: 1}}, - [OpenAI::BooleanModel, true] => true, - [OpenAI::BooleanModel, "true"] => "true", + [OpenAI::Internal::Type::BooleanModel, true] => true, + [OpenAI::Internal::Type::BooleanModel, "true"] => "true", [Integer, "1"] => "1", [Float, 1] => 1, [String, "one"] => "one", @@ -126,27 +126,27 @@ def test_coerce_errors class OpenAI::Test::EnumModelTest < Minitest::Test module E1 - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum TRUE = true end module E2 - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ONE = 1 TWO = 2 end module E3 - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ONE = 1.0 TWO = 2.0 end module E4 - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum ONE = :one TWO = :two @@ -216,14 +216,14 @@ def test_dump end class OpenAI::Test::CollectionModelTest < Minitest::Test - A1 = OpenAI::ArrayOf[-> { Integer }] - H1 = OpenAI::HashOf[Integer] + A1 = OpenAI::Internal::Type::ArrayOf[-> { Integer }] + H1 = OpenAI::Internal::Type::HashOf[Integer] - A2 = OpenAI::ArrayOf[H1] - H2 = OpenAI::HashOf[-> { A1 }] + A2 = OpenAI::Internal::Type::ArrayOf[H1] + H2 = OpenAI::Internal::Type::HashOf[-> { A1 }] - A3 = OpenAI::ArrayOf[Integer, nil?: true] - H3 = OpenAI::HashOf[Integer, nil?: true] + A3 = OpenAI::Internal::Type::ArrayOf[Integer, nil?: true] + H3 = OpenAI::Internal::Type::HashOf[Integer, nil?: true] def test_coerce cases = { @@ -263,7 +263,7 @@ def test_coerce end class OpenAI::Test::BaseModelTest < Minitest::Test - class M1 < OpenAI::BaseModel + class M1 < OpenAI::Internal::Type::BaseModel required :a, Integer end @@ -273,7 +273,7 @@ class M2 < M1 optional :c, String end - class M3 < OpenAI::BaseModel + class M3 < OpenAI::Internal::Type::BaseModel optional :c, const: :c required :d, const: :d end @@ -290,7 +290,7 @@ class M4 < M1 end end - class M5 < OpenAI::BaseModel + class M5 < OpenAI::Internal::Type::BaseModel request_only do required :c, const: :c end @@ -301,7 +301,7 @@ class M5 < OpenAI::BaseModel end class M6 < M1 - required :a, OpenAI::ArrayOf[M6] + required :a, OpenAI::Internal::Type::ArrayOf[M6] end def test_coerce @@ -337,7 +337,7 @@ def test_coerce assert_pattern do coerced = OpenAI::Internal::Type::Converter.coerce(target, input, state: state) assert_equal(coerced, coerced) - if coerced.is_a?(OpenAI::BaseModel) + if coerced.is_a?(OpenAI::Internal::Type::BaseModel) coerced.to_h => ^expect else coerced => ^expect @@ -403,27 +403,27 @@ def test_accessors class OpenAI::Test::UnionTest < Minitest::Test module U0 - extend OpenAI::Union + extend OpenAI::Internal::Type::Union end module U1 - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant const: :a variant const: 2 end - class M1 < OpenAI::BaseModel + class M1 < OpenAI::Internal::Type::BaseModel required :t, const: :a, api_name: :type optional :c, String end - class M2 < OpenAI::BaseModel + class M2 < OpenAI::Internal::Type::BaseModel required :type, const: :b optional :c, String end module U2 - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type variant :a, M1 @@ -431,7 +431,7 @@ module U2 end module U3 - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type variant :a, M1 @@ -439,37 +439,37 @@ module U3 end module U4 - extend OpenAI::Union + extend OpenAI::Internal::Type::Union discriminator :type variant String variant :a, M1 end - class M3 < OpenAI::BaseModel + class M3 < OpenAI::Internal::Type::BaseModel optional :recur, -> { U5 } required :a, Integer end - class M4 < OpenAI::BaseModel + class M4 < OpenAI::Internal::Type::BaseModel optional :recur, -> { U5 } - required :a, OpenAI::ArrayOf[-> { U5 }] + required :a, OpenAI::Internal::Type::ArrayOf[-> { U5 }] end - class M5 < OpenAI::BaseModel + class M5 < OpenAI::Internal::Type::BaseModel optional :recur, -> { U5 } - required :b, OpenAI::ArrayOf[-> { U5 }] + required :b, OpenAI::Internal::Type::ArrayOf[-> { U5 }] end module U5 - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant -> { M3 } variant -> { M4 } end module U6 - extend OpenAI::Union + extend OpenAI::Internal::Type::Union variant -> { M3 } variant -> { M5 } @@ -480,7 +480,7 @@ def test_accessors tap do model.recur flunk - rescue OpenAI::ConversionError => e + rescue OpenAI::Errors::ConversionError => e assert_kind_of(ArgumentError, e.cause) end end @@ -513,7 +513,7 @@ def test_coerce assert_pattern do coerced = OpenAI::Internal::Type::Converter.coerce(target, input, state: state) assert_equal(coerced, coerced) - if coerced.is_a?(OpenAI::BaseModel) + if coerced.is_a?(OpenAI::Internal::Type::BaseModel) coerced.to_h => ^expect else coerced => ^expect @@ -527,29 +527,29 @@ def test_coerce class OpenAI::Test::BaseModelQoLTest < Minitest::Test module E1 - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum A = 1 end module E2 - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum A = 1 end module E3 - extend OpenAI::Enum + extend OpenAI::Internal::Type::Enum A = 2 B = 3 end - class M1 < OpenAI::BaseModel + class M1 < OpenAI::Internal::Type::BaseModel required :a, Integer end - class M2 < OpenAI::BaseModel + class M2 < OpenAI::Internal::Type::BaseModel required :a, Integer, nil?: true end @@ -559,9 +559,9 @@ class M3 < M2 def test_equality cases = { - [OpenAI::Unknown, OpenAI::Unknown] => true, - [OpenAI::BooleanModel, OpenAI::BooleanModel] => true, - [OpenAI::Unknown, OpenAI::BooleanModel] => false, + [OpenAI::Internal::Type::Unknown, OpenAI::Internal::Type::Unknown] => true, + [OpenAI::Internal::Type::BooleanModel, OpenAI::Internal::Type::BooleanModel] => true, + [OpenAI::Internal::Type::Unknown, OpenAI::Internal::Type::BooleanModel] => false, [E1, E2] => true, [E1, E3] => false, [M1, M2] => false, diff --git a/test/openai/internal/util_test.rb b/test/openai/internal/util_test.rb index ed50d991..71ba50ed 100644 --- a/test/openai/internal/util_test.rb +++ b/test/openai/internal/util_test.rb @@ -42,7 +42,7 @@ def test_nested_left_map def test_omission merged = OpenAI::Internal::Util.deep_merge( {b: {b2: 1, b3: {c: 4, d: 5}}}, - {b: {b2: 1, b3: {c: OpenAI::Internal::Util::OMIT, d: 5}}} + {b: {b2: 1, b3: {c: OpenAI::Internal::OMIT, d: 5}}} ) assert_pattern do diff --git a/test/openai/resources/batches_test.rb b/test/openai/resources/batches_test.rb index 76c25672..8ba0d03e 100644 --- a/test/openai/resources/batches_test.rb +++ b/test/openai/resources/batches_test.rb @@ -34,7 +34,7 @@ def test_create_required_params failed_at: Integer | nil, finalizing_at: Integer | nil, in_progress_at: Integer | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, output_file_id: String | nil, request_counts: OpenAI::Models::BatchRequestCounts | nil } @@ -67,7 +67,7 @@ def test_retrieve failed_at: Integer | nil, finalizing_at: Integer | nil, in_progress_at: Integer | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, output_file_id: String | nil, request_counts: OpenAI::Models::BatchRequestCounts | nil } @@ -107,7 +107,7 @@ def test_list failed_at: Integer | nil, finalizing_at: Integer | nil, in_progress_at: Integer | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, output_file_id: String | nil, request_counts: OpenAI::Models::BatchRequestCounts | nil } @@ -140,7 +140,7 @@ def test_cancel failed_at: Integer | nil, finalizing_at: Integer | nil, in_progress_at: Integer | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, output_file_id: String | nil, request_counts: OpenAI::Models::BatchRequestCounts | nil } diff --git a/test/openai/resources/beta/assistants_test.rb b/test/openai/resources/beta/assistants_test.rb index f618bea7..0e89f80e 100644 --- a/test/openai/resources/beta/assistants_test.rb +++ b/test/openai/resources/beta/assistants_test.rb @@ -16,11 +16,11 @@ def test_create_required_params created_at: Integer, description: String | nil, instructions: String | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, model: String, name: String | nil, object: Symbol, - tools: ^(OpenAI::ArrayOf[union: OpenAI::Models::Beta::AssistantTool]), + tools: ^(OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::AssistantTool]), response_format: OpenAI::Models::Beta::AssistantResponseFormatOption | nil, temperature: Float | nil, tool_resources: OpenAI::Models::Beta::Assistant::ToolResources | nil, @@ -42,11 +42,11 @@ def test_retrieve created_at: Integer, description: String | nil, instructions: String | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, model: String, name: String | nil, object: Symbol, - tools: ^(OpenAI::ArrayOf[union: OpenAI::Models::Beta::AssistantTool]), + tools: ^(OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::AssistantTool]), response_format: OpenAI::Models::Beta::AssistantResponseFormatOption | nil, temperature: Float | nil, tool_resources: OpenAI::Models::Beta::Assistant::ToolResources | nil, @@ -68,11 +68,11 @@ def test_update created_at: Integer, description: String | nil, instructions: String | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, model: String, name: String | nil, object: Symbol, - tools: ^(OpenAI::ArrayOf[union: OpenAI::Models::Beta::AssistantTool]), + tools: ^(OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::AssistantTool]), response_format: OpenAI::Models::Beta::AssistantResponseFormatOption | nil, temperature: Float | nil, tool_resources: OpenAI::Models::Beta::Assistant::ToolResources | nil, @@ -101,11 +101,11 @@ def test_list created_at: Integer, description: String | nil, instructions: String | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, model: String, name: String | nil, object: Symbol, - tools: ^(OpenAI::ArrayOf[union: OpenAI::Models::Beta::AssistantTool]), + tools: ^(OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::AssistantTool]), response_format: OpenAI::Models::Beta::AssistantResponseFormatOption | nil, temperature: Float | nil, tool_resources: OpenAI::Models::Beta::Assistant::ToolResources | nil, @@ -124,7 +124,7 @@ def test_delete assert_pattern do response => { id: String, - deleted: OpenAI::BooleanModel, + deleted: OpenAI::Internal::Type::BooleanModel, object: Symbol } end diff --git a/test/openai/resources/beta/threads/messages_test.rb b/test/openai/resources/beta/threads/messages_test.rb index 7a8fdc90..5dec984b 100644 --- a/test/openai/resources/beta/threads/messages_test.rb +++ b/test/openai/resources/beta/threads/messages_test.rb @@ -14,13 +14,13 @@ def test_create_required_params response => { id: String, assistant_id: String | nil, - attachments: ^(OpenAI::ArrayOf[OpenAI::Models::Beta::Threads::Message::Attachment]) | nil, + attachments: ^(OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Beta::Threads::Message::Attachment]) | nil, completed_at: Integer | nil, - content: ^(OpenAI::ArrayOf[union: OpenAI::Models::Beta::Threads::MessageContent]), + content: ^(OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::Threads::MessageContent]), created_at: Integer, incomplete_at: Integer | nil, incomplete_details: OpenAI::Models::Beta::Threads::Message::IncompleteDetails | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, object: Symbol, role: OpenAI::Models::Beta::Threads::Message::Role, run_id: String | nil, @@ -41,13 +41,13 @@ def test_retrieve_required_params response => { id: String, assistant_id: String | nil, - attachments: ^(OpenAI::ArrayOf[OpenAI::Models::Beta::Threads::Message::Attachment]) | nil, + attachments: ^(OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Beta::Threads::Message::Attachment]) | nil, completed_at: Integer | nil, - content: ^(OpenAI::ArrayOf[union: OpenAI::Models::Beta::Threads::MessageContent]), + content: ^(OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::Threads::MessageContent]), created_at: Integer, incomplete_at: Integer | nil, incomplete_details: OpenAI::Models::Beta::Threads::Message::IncompleteDetails | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, object: Symbol, role: OpenAI::Models::Beta::Threads::Message::Role, run_id: String | nil, @@ -68,13 +68,13 @@ def test_update_required_params response => { id: String, assistant_id: String | nil, - attachments: ^(OpenAI::ArrayOf[OpenAI::Models::Beta::Threads::Message::Attachment]) | nil, + attachments: ^(OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Beta::Threads::Message::Attachment]) | nil, completed_at: Integer | nil, - content: ^(OpenAI::ArrayOf[union: OpenAI::Models::Beta::Threads::MessageContent]), + content: ^(OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::Threads::MessageContent]), created_at: Integer, incomplete_at: Integer | nil, incomplete_details: OpenAI::Models::Beta::Threads::Message::IncompleteDetails | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, object: Symbol, role: OpenAI::Models::Beta::Threads::Message::Role, run_id: String | nil, @@ -102,13 +102,13 @@ def test_list row => { id: String, assistant_id: String | nil, - attachments: ^(OpenAI::ArrayOf[OpenAI::Models::Beta::Threads::Message::Attachment]) | nil, + attachments: ^(OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Beta::Threads::Message::Attachment]) | nil, completed_at: Integer | nil, - content: ^(OpenAI::ArrayOf[union: OpenAI::Models::Beta::Threads::MessageContent]), + content: ^(OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::Threads::MessageContent]), created_at: Integer, incomplete_at: Integer | nil, incomplete_details: OpenAI::Models::Beta::Threads::Message::IncompleteDetails | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, object: Symbol, role: OpenAI::Models::Beta::Threads::Message::Role, run_id: String | nil, @@ -128,7 +128,7 @@ def test_delete_required_params assert_pattern do response => { id: String, - deleted: OpenAI::BooleanModel, + deleted: OpenAI::Internal::Type::BooleanModel, object: Symbol } end diff --git a/test/openai/resources/beta/threads/runs/steps_test.rb b/test/openai/resources/beta/threads/runs/steps_test.rb index c05de77f..200fc186 100644 --- a/test/openai/resources/beta/threads/runs/steps_test.rb +++ b/test/openai/resources/beta/threads/runs/steps_test.rb @@ -20,7 +20,7 @@ def test_retrieve_required_params expired_at: Integer | nil, failed_at: Integer | nil, last_error: OpenAI::Models::Beta::Threads::Runs::RunStep::LastError | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, object: Symbol, run_id: String, status: OpenAI::Models::Beta::Threads::Runs::RunStep::Status, @@ -56,7 +56,7 @@ def test_list_required_params expired_at: Integer | nil, failed_at: Integer | nil, last_error: OpenAI::Models::Beta::Threads::Runs::RunStep::LastError | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, object: Symbol, run_id: String, status: OpenAI::Models::Beta::Threads::Runs::RunStep::Status, diff --git a/test/openai/resources/beta/threads/runs_test.rb b/test/openai/resources/beta/threads/runs_test.rb index 1147966b..749bb638 100644 --- a/test/openai/resources/beta/threads/runs_test.rb +++ b/test/openai/resources/beta/threads/runs_test.rb @@ -24,17 +24,17 @@ def test_create_required_params last_error: OpenAI::Models::Beta::Threads::Run::LastError | nil, max_completion_tokens: Integer | nil, max_prompt_tokens: Integer | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, model: String, object: Symbol, - parallel_tool_calls: OpenAI::BooleanModel, + parallel_tool_calls: OpenAI::Internal::Type::BooleanModel, required_action: OpenAI::Models::Beta::Threads::Run::RequiredAction | nil, response_format: OpenAI::Models::Beta::AssistantResponseFormatOption | nil, started_at: Integer | nil, status: OpenAI::Models::Beta::Threads::RunStatus, thread_id: String, tool_choice: OpenAI::Models::Beta::AssistantToolChoiceOption | nil, - tools: ^(OpenAI::ArrayOf[union: OpenAI::Models::Beta::AssistantTool]), + tools: ^(OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::AssistantTool]), truncation_strategy: OpenAI::Models::Beta::Threads::Run::TruncationStrategy | nil, usage: OpenAI::Models::Beta::Threads::Run::Usage | nil, temperature: Float | nil, @@ -64,17 +64,17 @@ def test_retrieve_required_params last_error: OpenAI::Models::Beta::Threads::Run::LastError | nil, max_completion_tokens: Integer | nil, max_prompt_tokens: Integer | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, model: String, object: Symbol, - parallel_tool_calls: OpenAI::BooleanModel, + parallel_tool_calls: OpenAI::Internal::Type::BooleanModel, required_action: OpenAI::Models::Beta::Threads::Run::RequiredAction | nil, response_format: OpenAI::Models::Beta::AssistantResponseFormatOption | nil, started_at: Integer | nil, status: OpenAI::Models::Beta::Threads::RunStatus, thread_id: String, tool_choice: OpenAI::Models::Beta::AssistantToolChoiceOption | nil, - tools: ^(OpenAI::ArrayOf[union: OpenAI::Models::Beta::AssistantTool]), + tools: ^(OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::AssistantTool]), truncation_strategy: OpenAI::Models::Beta::Threads::Run::TruncationStrategy | nil, usage: OpenAI::Models::Beta::Threads::Run::Usage | nil, temperature: Float | nil, @@ -104,17 +104,17 @@ def test_update_required_params last_error: OpenAI::Models::Beta::Threads::Run::LastError | nil, max_completion_tokens: Integer | nil, max_prompt_tokens: Integer | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, model: String, object: Symbol, - parallel_tool_calls: OpenAI::BooleanModel, + parallel_tool_calls: OpenAI::Internal::Type::BooleanModel, required_action: OpenAI::Models::Beta::Threads::Run::RequiredAction | nil, response_format: OpenAI::Models::Beta::AssistantResponseFormatOption | nil, started_at: Integer | nil, status: OpenAI::Models::Beta::Threads::RunStatus, thread_id: String, tool_choice: OpenAI::Models::Beta::AssistantToolChoiceOption | nil, - tools: ^(OpenAI::ArrayOf[union: OpenAI::Models::Beta::AssistantTool]), + tools: ^(OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::AssistantTool]), truncation_strategy: OpenAI::Models::Beta::Threads::Run::TruncationStrategy | nil, usage: OpenAI::Models::Beta::Threads::Run::Usage | nil, temperature: Float | nil, @@ -151,17 +151,17 @@ def test_list last_error: OpenAI::Models::Beta::Threads::Run::LastError | nil, max_completion_tokens: Integer | nil, max_prompt_tokens: Integer | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, model: String, object: Symbol, - parallel_tool_calls: OpenAI::BooleanModel, + parallel_tool_calls: OpenAI::Internal::Type::BooleanModel, required_action: OpenAI::Models::Beta::Threads::Run::RequiredAction | nil, response_format: OpenAI::Models::Beta::AssistantResponseFormatOption | nil, started_at: Integer | nil, status: OpenAI::Models::Beta::Threads::RunStatus, thread_id: String, tool_choice: OpenAI::Models::Beta::AssistantToolChoiceOption | nil, - tools: ^(OpenAI::ArrayOf[union: OpenAI::Models::Beta::AssistantTool]), + tools: ^(OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::AssistantTool]), truncation_strategy: OpenAI::Models::Beta::Threads::Run::TruncationStrategy | nil, usage: OpenAI::Models::Beta::Threads::Run::Usage | nil, temperature: Float | nil, @@ -191,17 +191,17 @@ def test_cancel_required_params last_error: OpenAI::Models::Beta::Threads::Run::LastError | nil, max_completion_tokens: Integer | nil, max_prompt_tokens: Integer | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, model: String, object: Symbol, - parallel_tool_calls: OpenAI::BooleanModel, + parallel_tool_calls: OpenAI::Internal::Type::BooleanModel, required_action: OpenAI::Models::Beta::Threads::Run::RequiredAction | nil, response_format: OpenAI::Models::Beta::AssistantResponseFormatOption | nil, started_at: Integer | nil, status: OpenAI::Models::Beta::Threads::RunStatus, thread_id: String, tool_choice: OpenAI::Models::Beta::AssistantToolChoiceOption | nil, - tools: ^(OpenAI::ArrayOf[union: OpenAI::Models::Beta::AssistantTool]), + tools: ^(OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::AssistantTool]), truncation_strategy: OpenAI::Models::Beta::Threads::Run::TruncationStrategy | nil, usage: OpenAI::Models::Beta::Threads::Run::Usage | nil, temperature: Float | nil, @@ -232,17 +232,17 @@ def test_submit_tool_outputs_required_params last_error: OpenAI::Models::Beta::Threads::Run::LastError | nil, max_completion_tokens: Integer | nil, max_prompt_tokens: Integer | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, model: String, object: Symbol, - parallel_tool_calls: OpenAI::BooleanModel, + parallel_tool_calls: OpenAI::Internal::Type::BooleanModel, required_action: OpenAI::Models::Beta::Threads::Run::RequiredAction | nil, response_format: OpenAI::Models::Beta::AssistantResponseFormatOption | nil, started_at: Integer | nil, status: OpenAI::Models::Beta::Threads::RunStatus, thread_id: String, tool_choice: OpenAI::Models::Beta::AssistantToolChoiceOption | nil, - tools: ^(OpenAI::ArrayOf[union: OpenAI::Models::Beta::AssistantTool]), + tools: ^(OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::AssistantTool]), truncation_strategy: OpenAI::Models::Beta::Threads::Run::TruncationStrategy | nil, usage: OpenAI::Models::Beta::Threads::Run::Usage | nil, temperature: Float | nil, diff --git a/test/openai/resources/beta/threads_test.rb b/test/openai/resources/beta/threads_test.rb index f7f1021f..f702b5ce 100644 --- a/test/openai/resources/beta/threads_test.rb +++ b/test/openai/resources/beta/threads_test.rb @@ -14,7 +14,7 @@ def test_create response => { id: String, created_at: Integer, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, object: Symbol, tool_resources: OpenAI::Models::Beta::Thread::ToolResources | nil } @@ -32,7 +32,7 @@ def test_retrieve response => { id: String, created_at: Integer, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, object: Symbol, tool_resources: OpenAI::Models::Beta::Thread::ToolResources | nil } @@ -50,7 +50,7 @@ def test_update response => { id: String, created_at: Integer, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, object: Symbol, tool_resources: OpenAI::Models::Beta::Thread::ToolResources | nil } @@ -67,7 +67,7 @@ def test_delete assert_pattern do response => { id: String, - deleted: OpenAI::BooleanModel, + deleted: OpenAI::Internal::Type::BooleanModel, object: Symbol } end @@ -94,17 +94,17 @@ def test_create_and_run_required_params last_error: OpenAI::Models::Beta::Threads::Run::LastError | nil, max_completion_tokens: Integer | nil, max_prompt_tokens: Integer | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, model: String, object: Symbol, - parallel_tool_calls: OpenAI::BooleanModel, + parallel_tool_calls: OpenAI::Internal::Type::BooleanModel, required_action: OpenAI::Models::Beta::Threads::Run::RequiredAction | nil, response_format: OpenAI::Models::Beta::AssistantResponseFormatOption | nil, started_at: Integer | nil, status: OpenAI::Models::Beta::Threads::RunStatus, thread_id: String, tool_choice: OpenAI::Models::Beta::AssistantToolChoiceOption | nil, - tools: ^(OpenAI::ArrayOf[union: OpenAI::Models::Beta::AssistantTool]), + tools: ^(OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Beta::AssistantTool]), truncation_strategy: OpenAI::Models::Beta::Threads::Run::TruncationStrategy | nil, usage: OpenAI::Models::Beta::Threads::Run::Usage | nil, temperature: Float | nil, diff --git a/test/openai/resources/chat/completions_test.rb b/test/openai/resources/chat/completions_test.rb index 00d91a0a..8eab790c 100644 --- a/test/openai/resources/chat/completions_test.rb +++ b/test/openai/resources/chat/completions_test.rb @@ -14,7 +14,7 @@ def test_create_required_params assert_pattern do response => { id: String, - choices: ^(OpenAI::ArrayOf[OpenAI::Models::Chat::ChatCompletion::Choice]), + choices: ^(OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Chat::ChatCompletion::Choice]), created: Integer, model: String, object: Symbol, @@ -35,7 +35,7 @@ def test_retrieve assert_pattern do response => { id: String, - choices: ^(OpenAI::ArrayOf[OpenAI::Models::Chat::ChatCompletion::Choice]), + choices: ^(OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Chat::ChatCompletion::Choice]), created: Integer, model: String, object: Symbol, @@ -56,7 +56,7 @@ def test_update_required_params assert_pattern do response => { id: String, - choices: ^(OpenAI::ArrayOf[OpenAI::Models::Chat::ChatCompletion::Choice]), + choices: ^(OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Chat::ChatCompletion::Choice]), created: Integer, model: String, object: Symbol, @@ -84,7 +84,7 @@ def test_list assert_pattern do row => { id: String, - choices: ^(OpenAI::ArrayOf[OpenAI::Models::Chat::ChatCompletion::Choice]), + choices: ^(OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Chat::ChatCompletion::Choice]), created: Integer, model: String, object: Symbol, @@ -105,7 +105,7 @@ def test_delete assert_pattern do response => { id: String, - deleted: OpenAI::BooleanModel, + deleted: OpenAI::Internal::Type::BooleanModel, object: Symbol } end diff --git a/test/openai/resources/completions_test.rb b/test/openai/resources/completions_test.rb index f7d94ab8..e37c502a 100644 --- a/test/openai/resources/completions_test.rb +++ b/test/openai/resources/completions_test.rb @@ -13,7 +13,7 @@ def test_create_required_params assert_pattern do response => { id: String, - choices: ^(OpenAI::ArrayOf[OpenAI::Models::CompletionChoice]), + choices: ^(OpenAI::Internal::Type::ArrayOf[OpenAI::Models::CompletionChoice]), created: Integer, model: String, object: Symbol, diff --git a/test/openai/resources/embeddings_test.rb b/test/openai/resources/embeddings_test.rb index d9eb182b..a280c031 100644 --- a/test/openai/resources/embeddings_test.rb +++ b/test/openai/resources/embeddings_test.rb @@ -16,7 +16,7 @@ def test_create_required_params assert_pattern do response => { - data: ^(OpenAI::ArrayOf[OpenAI::Models::Embedding]), + data: ^(OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Embedding]), model: String, object: Symbol, usage: OpenAI::Models::CreateEmbeddingResponse::Usage diff --git a/test/openai/resources/files_test.rb b/test/openai/resources/files_test.rb index c4a14255..78bfa904 100644 --- a/test/openai/resources/files_test.rb +++ b/test/openai/resources/files_test.rb @@ -86,7 +86,7 @@ def test_delete assert_pattern do response => { id: String, - deleted: OpenAI::BooleanModel, + deleted: OpenAI::Internal::Type::BooleanModel, object: Symbol } end diff --git a/test/openai/resources/fine_tuning/jobs_test.rb b/test/openai/resources/fine_tuning/jobs_test.rb index a953af3d..9f164929 100644 --- a/test/openai/resources/fine_tuning/jobs_test.rb +++ b/test/openai/resources/fine_tuning/jobs_test.rb @@ -21,15 +21,15 @@ def test_create_required_params model: String, object: Symbol, organization_id: String, - result_files: ^(OpenAI::ArrayOf[String]), + result_files: ^(OpenAI::Internal::Type::ArrayOf[String]), seed: Integer, status: OpenAI::Models::FineTuning::FineTuningJob::Status, trained_tokens: Integer | nil, training_file: String, validation_file: String | nil, estimated_finish: Integer | nil, - integrations: ^(OpenAI::ArrayOf[OpenAI::Models::FineTuning::FineTuningJobWandbIntegrationObject]) | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + integrations: ^(OpenAI::Internal::Type::ArrayOf[OpenAI::Models::FineTuning::FineTuningJobWandbIntegrationObject]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, method_: OpenAI::Models::FineTuning::FineTuningJob::Method | nil } end @@ -53,15 +53,15 @@ def test_retrieve model: String, object: Symbol, organization_id: String, - result_files: ^(OpenAI::ArrayOf[String]), + result_files: ^(OpenAI::Internal::Type::ArrayOf[String]), seed: Integer, status: OpenAI::Models::FineTuning::FineTuningJob::Status, trained_tokens: Integer | nil, training_file: String, validation_file: String | nil, estimated_finish: Integer | nil, - integrations: ^(OpenAI::ArrayOf[OpenAI::Models::FineTuning::FineTuningJobWandbIntegrationObject]) | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + integrations: ^(OpenAI::Internal::Type::ArrayOf[OpenAI::Models::FineTuning::FineTuningJobWandbIntegrationObject]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, method_: OpenAI::Models::FineTuning::FineTuningJob::Method | nil } end @@ -92,15 +92,15 @@ def test_list model: String, object: Symbol, organization_id: String, - result_files: ^(OpenAI::ArrayOf[String]), + result_files: ^(OpenAI::Internal::Type::ArrayOf[String]), seed: Integer, status: OpenAI::Models::FineTuning::FineTuningJob::Status, trained_tokens: Integer | nil, training_file: String, validation_file: String | nil, estimated_finish: Integer | nil, - integrations: ^(OpenAI::ArrayOf[OpenAI::Models::FineTuning::FineTuningJobWandbIntegrationObject]) | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + integrations: ^(OpenAI::Internal::Type::ArrayOf[OpenAI::Models::FineTuning::FineTuningJobWandbIntegrationObject]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, method_: OpenAI::Models::FineTuning::FineTuningJob::Method | nil } end @@ -124,15 +124,15 @@ def test_cancel model: String, object: Symbol, organization_id: String, - result_files: ^(OpenAI::ArrayOf[String]), + result_files: ^(OpenAI::Internal::Type::ArrayOf[String]), seed: Integer, status: OpenAI::Models::FineTuning::FineTuningJob::Status, trained_tokens: Integer | nil, training_file: String, validation_file: String | nil, estimated_finish: Integer | nil, - integrations: ^(OpenAI::ArrayOf[OpenAI::Models::FineTuning::FineTuningJobWandbIntegrationObject]) | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + integrations: ^(OpenAI::Internal::Type::ArrayOf[OpenAI::Models::FineTuning::FineTuningJobWandbIntegrationObject]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, method_: OpenAI::Models::FineTuning::FineTuningJob::Method | nil } end @@ -159,7 +159,7 @@ def test_list_events level: OpenAI::Models::FineTuning::FineTuningJobEvent::Level, message: String, object: Symbol, - data: OpenAI::Unknown | nil, + data: OpenAI::Internal::Type::Unknown | nil, type: OpenAI::Models::FineTuning::FineTuningJobEvent::Type | nil } end diff --git a/test/openai/resources/images_test.rb b/test/openai/resources/images_test.rb index c935d0c6..a9bb70a4 100644 --- a/test/openai/resources/images_test.rb +++ b/test/openai/resources/images_test.rb @@ -13,7 +13,7 @@ def test_create_variation_required_params assert_pattern do response => { created: Integer, - data: ^(OpenAI::ArrayOf[OpenAI::Models::Image]) + data: ^(OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Image]) } end end @@ -32,7 +32,7 @@ def test_edit_required_params assert_pattern do response => { created: Integer, - data: ^(OpenAI::ArrayOf[OpenAI::Models::Image]) + data: ^(OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Image]) } end end @@ -47,7 +47,7 @@ def test_generate_required_params assert_pattern do response => { created: Integer, - data: ^(OpenAI::ArrayOf[OpenAI::Models::Image]) + data: ^(OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Image]) } end end diff --git a/test/openai/resources/models_test.rb b/test/openai/resources/models_test.rb index 6316652e..06dac749 100644 --- a/test/openai/resources/models_test.rb +++ b/test/openai/resources/models_test.rb @@ -54,7 +54,7 @@ def test_delete assert_pattern do response => { id: String, - deleted: OpenAI::BooleanModel, + deleted: OpenAI::Internal::Type::BooleanModel, object: String } end diff --git a/test/openai/resources/moderations_test.rb b/test/openai/resources/moderations_test.rb index cbd3c69f..d77f7d0b 100644 --- a/test/openai/resources/moderations_test.rb +++ b/test/openai/resources/moderations_test.rb @@ -14,7 +14,7 @@ def test_create_required_params response => { id: String, model: String, - results: ^(OpenAI::ArrayOf[OpenAI::Models::Moderation]) + results: ^(OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Moderation]) } end end diff --git a/test/openai/resources/responses/input_items_test.rb b/test/openai/resources/responses/input_items_test.rb index 59a1cc5e..705de078 100644 --- a/test/openai/resources/responses/input_items_test.rb +++ b/test/openai/resources/responses/input_items_test.rb @@ -35,30 +35,30 @@ def test_list in { type: :message, id: String, - content: ^(OpenAI::ArrayOf[union: OpenAI::Models::Responses::ResponseInputContent]), + content: ^(OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Responses::ResponseInputContent]), role: OpenAI::Models::Responses::ResponseInputMessageItem::Role, status: OpenAI::Models::Responses::ResponseInputMessageItem::Status | nil } in { type: :message, id: String, - content: ^(OpenAI::ArrayOf[union: OpenAI::Models::Responses::ResponseOutputMessage::Content]), + content: ^(OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Responses::ResponseOutputMessage::Content]), role: Symbol, status: OpenAI::Models::Responses::ResponseOutputMessage::Status } in { type: :file_search_call, id: String, - queries: ^(OpenAI::ArrayOf[String]), + queries: ^(OpenAI::Internal::Type::ArrayOf[String]), status: OpenAI::Models::Responses::ResponseFileSearchToolCall::Status, - results: ^(OpenAI::ArrayOf[OpenAI::Models::Responses::ResponseFileSearchToolCall::Result]) | nil + results: ^(OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Responses::ResponseFileSearchToolCall::Result]) | nil } in { type: :computer_call, id: String, action: OpenAI::Models::Responses::ResponseComputerToolCall::Action, call_id: String, - pending_safety_checks: ^(OpenAI::ArrayOf[OpenAI::Models::Responses::ResponseComputerToolCall::PendingSafetyCheck]), + pending_safety_checks: ^(OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Responses::ResponseComputerToolCall::PendingSafetyCheck]), status: OpenAI::Models::Responses::ResponseComputerToolCall::Status } in { @@ -66,7 +66,7 @@ def test_list id: String, call_id: String, output: OpenAI::Models::Responses::ResponseComputerToolCallOutputScreenshot, - acknowledged_safety_checks: ^(OpenAI::ArrayOf[OpenAI::Models::Responses::ResponseComputerToolCallOutputItem::AcknowledgedSafetyCheck]) | nil, + acknowledged_safety_checks: ^(OpenAI::Internal::Type::ArrayOf[OpenAI::Models::Responses::ResponseComputerToolCallOutputItem::AcknowledgedSafetyCheck]) | nil, status: OpenAI::Models::Responses::ResponseComputerToolCallOutputItem::Status | nil } in {type: :web_search_call, id: String, status: OpenAI::Models::Responses::ResponseFunctionWebSearch::Status} diff --git a/test/openai/resources/responses_test.rb b/test/openai/resources/responses_test.rb index d0e0157e..3b452659 100644 --- a/test/openai/resources/responses_test.rb +++ b/test/openai/resources/responses_test.rb @@ -17,14 +17,14 @@ def test_create_required_params error: OpenAI::Models::Responses::ResponseError | nil, incomplete_details: OpenAI::Models::Responses::Response::IncompleteDetails | nil, instructions: String | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, model: OpenAI::Models::ResponsesModel, object: Symbol, - output: ^(OpenAI::ArrayOf[union: OpenAI::Models::Responses::ResponseOutputItem]), - parallel_tool_calls: OpenAI::BooleanModel, + output: ^(OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Responses::ResponseOutputItem]), + parallel_tool_calls: OpenAI::Internal::Type::BooleanModel, temperature: Float | nil, tool_choice: OpenAI::Models::Responses::Response::ToolChoice, - tools: ^(OpenAI::ArrayOf[union: OpenAI::Models::Responses::Tool]), + tools: ^(OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Responses::Tool]), top_p: Float | nil, max_output_tokens: Integer | nil, previous_response_id: String | nil, @@ -52,14 +52,14 @@ def test_retrieve error: OpenAI::Models::Responses::ResponseError | nil, incomplete_details: OpenAI::Models::Responses::Response::IncompleteDetails | nil, instructions: String | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, model: OpenAI::Models::ResponsesModel, object: Symbol, - output: ^(OpenAI::ArrayOf[union: OpenAI::Models::Responses::ResponseOutputItem]), - parallel_tool_calls: OpenAI::BooleanModel, + output: ^(OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Responses::ResponseOutputItem]), + parallel_tool_calls: OpenAI::Internal::Type::BooleanModel, temperature: Float | nil, tool_choice: OpenAI::Models::Responses::Response::ToolChoice, - tools: ^(OpenAI::ArrayOf[union: OpenAI::Models::Responses::Tool]), + tools: ^(OpenAI::Internal::Type::ArrayOf[union: OpenAI::Models::Responses::Tool]), top_p: Float | nil, max_output_tokens: Integer | nil, previous_response_id: String | nil, diff --git a/test/openai/resources/vector_stores/file_batches_test.rb b/test/openai/resources/vector_stores/file_batches_test.rb index 538469fe..55dce57e 100644 --- a/test/openai/resources/vector_stores/file_batches_test.rb +++ b/test/openai/resources/vector_stores/file_batches_test.rb @@ -83,7 +83,7 @@ def test_list_files_required_params status: OpenAI::Models::VectorStores::VectorStoreFile::Status, usage_bytes: Integer, vector_store_id: String, - attributes: ^(OpenAI::HashOf[union: OpenAI::Models::VectorStores::VectorStoreFile::Attribute]) | nil, + attributes: ^(OpenAI::Internal::Type::HashOf[union: OpenAI::Models::VectorStores::VectorStoreFile::Attribute]) | nil, chunking_strategy: OpenAI::Models::FileChunkingStrategy | nil } end diff --git a/test/openai/resources/vector_stores/files_test.rb b/test/openai/resources/vector_stores/files_test.rb index 5d806b95..279fda94 100644 --- a/test/openai/resources/vector_stores/files_test.rb +++ b/test/openai/resources/vector_stores/files_test.rb @@ -19,7 +19,7 @@ def test_create_required_params status: OpenAI::Models::VectorStores::VectorStoreFile::Status, usage_bytes: Integer, vector_store_id: String, - attributes: ^(OpenAI::HashOf[union: OpenAI::Models::VectorStores::VectorStoreFile::Attribute]) | nil, + attributes: ^(OpenAI::Internal::Type::HashOf[union: OpenAI::Models::VectorStores::VectorStoreFile::Attribute]) | nil, chunking_strategy: OpenAI::Models::FileChunkingStrategy | nil } end @@ -41,7 +41,7 @@ def test_retrieve_required_params status: OpenAI::Models::VectorStores::VectorStoreFile::Status, usage_bytes: Integer, vector_store_id: String, - attributes: ^(OpenAI::HashOf[union: OpenAI::Models::VectorStores::VectorStoreFile::Attribute]) | nil, + attributes: ^(OpenAI::Internal::Type::HashOf[union: OpenAI::Models::VectorStores::VectorStoreFile::Attribute]) | nil, chunking_strategy: OpenAI::Models::FileChunkingStrategy | nil } end @@ -68,7 +68,7 @@ def test_update_required_params status: OpenAI::Models::VectorStores::VectorStoreFile::Status, usage_bytes: Integer, vector_store_id: String, - attributes: ^(OpenAI::HashOf[union: OpenAI::Models::VectorStores::VectorStoreFile::Attribute]) | nil, + attributes: ^(OpenAI::Internal::Type::HashOf[union: OpenAI::Models::VectorStores::VectorStoreFile::Attribute]) | nil, chunking_strategy: OpenAI::Models::FileChunkingStrategy | nil } end @@ -97,7 +97,7 @@ def test_list status: OpenAI::Models::VectorStores::VectorStoreFile::Status, usage_bytes: Integer, vector_store_id: String, - attributes: ^(OpenAI::HashOf[union: OpenAI::Models::VectorStores::VectorStoreFile::Attribute]) | nil, + attributes: ^(OpenAI::Internal::Type::HashOf[union: OpenAI::Models::VectorStores::VectorStoreFile::Attribute]) | nil, chunking_strategy: OpenAI::Models::FileChunkingStrategy | nil } end @@ -113,7 +113,7 @@ def test_delete_required_params assert_pattern do response => { id: String, - deleted: OpenAI::BooleanModel, + deleted: OpenAI::Internal::Type::BooleanModel, object: Symbol } end diff --git a/test/openai/resources/vector_stores_test.rb b/test/openai/resources/vector_stores_test.rb index c12e8413..1a45528c 100644 --- a/test/openai/resources/vector_stores_test.rb +++ b/test/openai/resources/vector_stores_test.rb @@ -16,7 +16,7 @@ def test_create created_at: Integer, file_counts: OpenAI::Models::VectorStore::FileCounts, last_active_at: Integer | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, name: String, object: Symbol, status: OpenAI::Models::VectorStore::Status, @@ -40,7 +40,7 @@ def test_retrieve created_at: Integer, file_counts: OpenAI::Models::VectorStore::FileCounts, last_active_at: Integer | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, name: String, object: Symbol, status: OpenAI::Models::VectorStore::Status, @@ -64,7 +64,7 @@ def test_update created_at: Integer, file_counts: OpenAI::Models::VectorStore::FileCounts, last_active_at: Integer | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, name: String, object: Symbol, status: OpenAI::Models::VectorStore::Status, @@ -95,7 +95,7 @@ def test_list created_at: Integer, file_counts: OpenAI::Models::VectorStore::FileCounts, last_active_at: Integer | nil, - metadata: ^(OpenAI::HashOf[String]) | nil, + metadata: ^(OpenAI::Internal::Type::HashOf[String]) | nil, name: String, object: Symbol, status: OpenAI::Models::VectorStore::Status, @@ -116,7 +116,7 @@ def test_delete assert_pattern do response => { id: String, - deleted: OpenAI::BooleanModel, + deleted: OpenAI::Internal::Type::BooleanModel, object: Symbol } end @@ -138,8 +138,8 @@ def test_search_required_params assert_pattern do row => { - attributes: ^(OpenAI::HashOf[union: OpenAI::Models::VectorStoreSearchResponse::Attribute]) | nil, - content: ^(OpenAI::ArrayOf[OpenAI::Models::VectorStoreSearchResponse::Content]), + attributes: ^(OpenAI::Internal::Type::HashOf[union: OpenAI::Models::VectorStoreSearchResponse::Attribute]) | nil, + content: ^(OpenAI::Internal::Type::ArrayOf[OpenAI::Models::VectorStoreSearchResponse::Content]), file_id: String, filename: String, score: Float