You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
# Some parameter documentations has been truncated, see
@@ -194,7 +196,7 @@ class ThreadCreateAndRunParams < OpenAI::Internal::Type::BaseModel
194
196
#
195
197
# @param top_p [Float, nil] An alternative to sampling with temperature, called nucleus sampling, where the
196
198
#
197
-
# @param truncation_strategy [OpenAI::Beta::TruncationObject, nil] Controls for how a thread will be truncated prior to the run. Use this to contro
199
+
# @param truncation_strategy [OpenAI::Beta::ThreadCreateAndRunParams::TruncationStrategy, nil] Controls for how a thread will be truncated prior to the run. Use this to contro
# The number of most recent messages from the thread when constructing the context
724
+
# for the run.
725
+
#
726
+
# @return [Integer, nil]
727
+
optional:last_messages,Integer,nil?: true
728
+
729
+
# @!method initialize(type:, last_messages: nil)
730
+
# Some parameter documentations has been truncated, see
731
+
# {OpenAI::Beta::ThreadCreateAndRunParams::TruncationStrategy} for more details.
732
+
#
733
+
# Controls for how a thread will be truncated prior to the run. Use this to
734
+
# control the intial context window of the run.
735
+
#
736
+
# @param type [Symbol, OpenAI::Beta::ThreadCreateAndRunParams::TruncationStrategy::Type] The truncation strategy to use for the thread. The default is `auto`. If set to
737
+
#
738
+
# @param last_messages [Integer, nil] The number of most recent messages from the thread when constructing the context
739
+
740
+
# The truncation strategy to use for the thread. The default is `auto`. If set to
741
+
# `last_messages`, the thread will be truncated to the n most recent messages in
742
+
# the thread. When set to `auto`, messages in the middle of the thread will be
743
+
# dropped to fit the context length of the model, `max_prompt_tokens`.
# Usage statistics related to the run. This value will be `null` if the run is not
@@ -270,7 +270,7 @@ class Run < OpenAI::Internal::Type::BaseModel
270
270
#
271
271
# @param tools [Array<OpenAI::Beta::CodeInterpreterTool, OpenAI::Beta::FileSearchTool, OpenAI::Beta::FunctionTool>] The list of tools that the [assistant](https://platform.openai.com/docs/api-refe
272
272
#
273
-
# @param truncation_strategy [OpenAI::Beta::TruncationObject, nil] Controls for how a thread will be truncated prior to the run. Use this to contro
273
+
# @param truncation_strategy [OpenAI::Beta::Threads::Run::TruncationStrategy, nil] Controls for how a thread will be truncated prior to the run. Use this to contro
274
274
#
275
275
# @param usage [OpenAI::Beta::Threads::Run::Usage, nil] Usage statistics related to the run. This value will be `null` if the run is not
276
276
#
@@ -392,6 +392,52 @@ class SubmitToolOutputs < OpenAI::Internal::Type::BaseModel
# The number of most recent messages from the thread when constructing the context
408
+
# for the run.
409
+
#
410
+
# @return [Integer, nil]
411
+
optional:last_messages,Integer,nil?: true
412
+
413
+
# @!method initialize(type:, last_messages: nil)
414
+
# Some parameter documentations has been truncated, see
415
+
# {OpenAI::Beta::Threads::Run::TruncationStrategy} for more details.
416
+
#
417
+
# Controls for how a thread will be truncated prior to the run. Use this to
418
+
# control the intial context window of the run.
419
+
#
420
+
# @param type [Symbol, OpenAI::Beta::Threads::Run::TruncationStrategy::Type] The truncation strategy to use for the thread. The default is `auto`. If set to
421
+
#
422
+
# @param last_messages [Integer, nil] The number of most recent messages from the thread when constructing the context
423
+
424
+
# The truncation strategy to use for the thread. The default is `auto`. If set to
425
+
# `last_messages`, the thread will be truncated to the n most recent messages in
426
+
# the thread. When set to `auto`, messages in the middle of the thread will be
427
+
# dropped to fit the context length of the model, `max_prompt_tokens`.
# Some parameter documentations has been truncated, see
@@ -226,7 +228,7 @@ class RunCreateParams < OpenAI::Internal::Type::BaseModel
226
228
#
227
229
# @param top_p [Float, nil] An alternative to sampling with temperature, called nucleus sampling, where the
228
230
#
229
-
# @param truncation_strategy [OpenAI::Beta::TruncationObject, nil] Controls for how a thread will be truncated prior to the run. Use this to contro
231
+
# @param truncation_strategy [OpenAI::Beta::Threads::RunCreateParams::TruncationStrategy, nil] Controls for how a thread will be truncated prior to the run. Use this to contro
# The number of most recent messages from the thread when constructing the context
435
+
# for the run.
436
+
#
437
+
# @return [Integer, nil]
438
+
optional:last_messages,Integer,nil?: true
439
+
440
+
# @!method initialize(type:, last_messages: nil)
441
+
# Some parameter documentations has been truncated, see
442
+
# {OpenAI::Beta::Threads::RunCreateParams::TruncationStrategy} for more details.
443
+
#
444
+
# Controls for how a thread will be truncated prior to the run. Use this to
445
+
# control the intial context window of the run.
446
+
#
447
+
# @param type [Symbol, OpenAI::Beta::Threads::RunCreateParams::TruncationStrategy::Type] The truncation strategy to use for the thread. The default is `auto`. If set to
448
+
#
449
+
# @param last_messages [Integer, nil] The number of most recent messages from the thread when constructing the context
450
+
451
+
# The truncation strategy to use for the thread. The default is `auto`. If set to
452
+
# `last_messages`, the thread will be truncated to the n most recent messages in
453
+
# the thread. When set to `auto`, messages in the middle of the thread will be
454
+
# dropped to fit the context length of the model, `max_prompt_tokens`.
0 commit comments