From aa445cab5c93c6908697fe98e73e16963330b141 Mon Sep 17 00:00:00 2001 From: giovx Date: Wed, 8 Oct 2025 11:30:46 +0200 Subject: [PATCH 1/3] chore(package): bump jiter to >=0.10.0 to support Python 3.14 (#2618) Co-authored-by: Robert Craigie --- pyproject.toml | 2 +- requirements-dev.lock | 2 +- requirements.lock | 2 +- 3 files changed, 3 insertions(+), 3 deletions(-) diff --git a/pyproject.toml b/pyproject.toml index 7e9f9bd1ee..bb5eb7cf41 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -15,7 +15,7 @@ dependencies = [ "distro>=1.7.0, <2", "sniffio", "tqdm > 4", - "jiter>=0.4.0, <1", + "jiter>=0.10.0, <1", ] requires-python = ">= 3.8" classifiers = [ diff --git a/requirements-dev.lock b/requirements-dev.lock index 0bd1c2c70f..e2446f2222 100644 --- a/requirements-dev.lock +++ b/requirements-dev.lock @@ -91,7 +91,7 @@ importlib-metadata==7.0.0 iniconfig==2.0.0 # via pytest inline-snapshot==0.28.0 -jiter==0.5.0 +jiter==0.11.0 # via openai markdown-it-py==3.0.0 # via rich diff --git a/requirements.lock b/requirements.lock index a2b6845942..aaca0834db 100644 --- a/requirements.lock +++ b/requirements.lock @@ -51,7 +51,7 @@ idna==3.4 # via anyio # via httpx # via yarl -jiter==0.6.1 +jiter==0.11.0 # via openai multidict==6.5.0 # via aiohttp From aa49f626a6ea9d77ad008badfb3741e16232d62f Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Fri, 10 Oct 2025 01:09:07 +0000 Subject: [PATCH 2/3] feat(api): comparison filter in/not in --- .stats.yml | 6 +++--- src/openai/resources/beta/assistants.py | 12 ++++++++++++ src/openai/resources/beta/threads/runs/runs.py | 18 ++++++++++++++++++ .../resources/chat/completions/completions.py | 18 ++++++++++++++++++ src/openai/resources/files.py | 4 ++-- .../types/beta/assistant_create_params.py | 3 +++ .../types/beta/assistant_update_params.py | 3 +++ .../types/beta/threads/run_create_params.py | 3 +++ .../types/chat/completion_create_params.py | 3 +++ .../create_eval_completions_run_data_source.py | 3 +++ ...e_eval_completions_run_data_source_param.py | 3 +++ src/openai/types/evals/run_cancel_response.py | 6 ++++++ src/openai/types/evals/run_create_params.py | 6 ++++++ src/openai/types/evals/run_create_response.py | 6 ++++++ src/openai/types/evals/run_list_response.py | 6 ++++++ .../types/evals/run_retrieve_response.py | 6 ++++++ src/openai/types/graders/score_model_grader.py | 3 +++ .../types/graders/score_model_grader_param.py | 3 +++ src/openai/types/shared/comparison_filter.py | 10 +++++++--- src/openai/types/shared/reasoning.py | 3 +++ .../types/shared_params/comparison_filter.py | 10 ++++++++-- src/openai/types/shared_params/reasoning.py | 3 +++ .../types/vector_stores/vector_store_file.py | 2 +- 23 files changed, 129 insertions(+), 11 deletions(-) diff --git a/.stats.yml b/.stats.yml index 2dc7422e7e..b5d9915ab8 100644 --- a/.stats.yml +++ b/.stats.yml @@ -1,4 +1,4 @@ configured_endpoints: 136 -openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/openai%2Fopenai-d64cf80d2ebddf175c5578f68226a3d5bbd3f7fd8d62ccac2205f3fc05a355ee.yml -openapi_spec_hash: d51e0d60d0c536f210b597a211bc5af0 -config_hash: e7c42016df9c6bd7bd6ff15101b9bc9b +openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/openai%2Fopenai-e66e85fb7f72477256dca1acb6b23396989d381c5c1b318de564195436bcb93f.yml +openapi_spec_hash: 0a4bbb5aa0ae532a072bd6b3854e70b1 +config_hash: 89bf7bb3a1f9439ffc6ea0e7dc57ba9b diff --git a/src/openai/resources/beta/assistants.py b/src/openai/resources/beta/assistants.py index ddac9a79cb..a958c0caa1 100644 --- a/src/openai/resources/beta/assistants.py +++ b/src/openai/resources/beta/assistants.py @@ -102,6 +102,9 @@ def create( effort can result in faster responses and fewer tokens used on reasoning in a response. + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. + response_format: Specifies the format that the model must output. Compatible with [GPT-4o](https://platform.openai.com/docs/models#gpt-4o), [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4), @@ -309,6 +312,9 @@ def update( effort can result in faster responses and fewer tokens used on reasoning in a response. + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. + response_format: Specifies the format that the model must output. Compatible with [GPT-4o](https://platform.openai.com/docs/models#gpt-4o), [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4), @@ -555,6 +561,9 @@ async def create( effort can result in faster responses and fewer tokens used on reasoning in a response. + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. + response_format: Specifies the format that the model must output. Compatible with [GPT-4o](https://platform.openai.com/docs/models#gpt-4o), [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4), @@ -762,6 +771,9 @@ async def update( effort can result in faster responses and fewer tokens used on reasoning in a response. + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. + response_format: Specifies the format that the model must output. Compatible with [GPT-4o](https://platform.openai.com/docs/models#gpt-4o), [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4), diff --git a/src/openai/resources/beta/threads/runs/runs.py b/src/openai/resources/beta/threads/runs/runs.py index ec2dfa84cd..2753f5817e 100644 --- a/src/openai/resources/beta/threads/runs/runs.py +++ b/src/openai/resources/beta/threads/runs/runs.py @@ -173,6 +173,9 @@ def create( effort can result in faster responses and fewer tokens used on reasoning in a response. + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. + response_format: Specifies the format that the model must output. Compatible with [GPT-4o](https://platform.openai.com/docs/models#gpt-4o), [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4), @@ -327,6 +330,9 @@ def create( effort can result in faster responses and fewer tokens used on reasoning in a response. + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. + response_format: Specifies the format that the model must output. Compatible with [GPT-4o](https://platform.openai.com/docs/models#gpt-4o), [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4), @@ -477,6 +483,9 @@ def create( effort can result in faster responses and fewer tokens used on reasoning in a response. + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. + response_format: Specifies the format that the model must output. Compatible with [GPT-4o](https://platform.openai.com/docs/models#gpt-4o), [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4), @@ -1603,6 +1612,9 @@ async def create( effort can result in faster responses and fewer tokens used on reasoning in a response. + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. + response_format: Specifies the format that the model must output. Compatible with [GPT-4o](https://platform.openai.com/docs/models#gpt-4o), [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4), @@ -1757,6 +1769,9 @@ async def create( effort can result in faster responses and fewer tokens used on reasoning in a response. + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. + response_format: Specifies the format that the model must output. Compatible with [GPT-4o](https://platform.openai.com/docs/models#gpt-4o), [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4), @@ -1907,6 +1922,9 @@ async def create( effort can result in faster responses and fewer tokens used on reasoning in a response. + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. + response_format: Specifies the format that the model must output. Compatible with [GPT-4o](https://platform.openai.com/docs/models#gpt-4o), [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4), diff --git a/src/openai/resources/chat/completions/completions.py b/src/openai/resources/chat/completions/completions.py index 329634ba43..4b73c69ae9 100644 --- a/src/openai/resources/chat/completions/completions.py +++ b/src/openai/resources/chat/completions/completions.py @@ -407,6 +407,9 @@ def create( effort can result in faster responses and fewer tokens used on reasoning in a response. + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. + response_format: An object specifying the format that the model must output. Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured @@ -704,6 +707,9 @@ def create( effort can result in faster responses and fewer tokens used on reasoning in a response. + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. + response_format: An object specifying the format that the model must output. Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured @@ -992,6 +998,9 @@ def create( effort can result in faster responses and fewer tokens used on reasoning in a response. + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. + response_format: An object specifying the format that the model must output. Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured @@ -1845,6 +1854,9 @@ async def create( effort can result in faster responses and fewer tokens used on reasoning in a response. + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. + response_format: An object specifying the format that the model must output. Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured @@ -2142,6 +2154,9 @@ async def create( effort can result in faster responses and fewer tokens used on reasoning in a response. + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. + response_format: An object specifying the format that the model must output. Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured @@ -2430,6 +2445,9 @@ async def create( effort can result in faster responses and fewer tokens used on reasoning in a response. + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. + response_format: An object specifying the format that the model must output. Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured diff --git a/src/openai/resources/files.py b/src/openai/resources/files.py index 77bb2d613c..bb9ad43b1b 100644 --- a/src/openai/resources/files.py +++ b/src/openai/resources/files.py @@ -236,7 +236,7 @@ def delete( timeout: float | httpx.Timeout | None | NotGiven = not_given, ) -> FileDeleted: """ - Delete a file. + Delete a file and remove it from all vector stores. Args: extra_headers: Send extra headers @@ -553,7 +553,7 @@ async def delete( timeout: float | httpx.Timeout | None | NotGiven = not_given, ) -> FileDeleted: """ - Delete a file. + Delete a file and remove it from all vector stores. Args: extra_headers: Send extra headers diff --git a/src/openai/types/beta/assistant_create_params.py b/src/openai/types/beta/assistant_create_params.py index 07f8f28f02..6fb1551fa5 100644 --- a/src/openai/types/beta/assistant_create_params.py +++ b/src/openai/types/beta/assistant_create_params.py @@ -65,6 +65,9 @@ class AssistantCreateParams(TypedDict, total=False): supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response. + + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. """ response_format: Optional[AssistantResponseFormatOptionParam] diff --git a/src/openai/types/beta/assistant_update_params.py b/src/openai/types/beta/assistant_update_params.py index 45d9f984b2..6d20b8e01f 100644 --- a/src/openai/types/beta/assistant_update_params.py +++ b/src/openai/types/beta/assistant_update_params.py @@ -100,6 +100,9 @@ class AssistantUpdateParams(TypedDict, total=False): supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response. + + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. """ response_format: Optional[AssistantResponseFormatOptionParam] diff --git a/src/openai/types/beta/threads/run_create_params.py b/src/openai/types/beta/threads/run_create_params.py index cfd272f5ad..3190c8b308 100644 --- a/src/openai/types/beta/threads/run_create_params.py +++ b/src/openai/types/beta/threads/run_create_params.py @@ -114,6 +114,9 @@ class RunCreateParamsBase(TypedDict, total=False): supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response. + + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. """ response_format: Optional[AssistantResponseFormatOptionParam] diff --git a/src/openai/types/chat/completion_create_params.py b/src/openai/types/chat/completion_create_params.py index 2ae81dfbc2..8b0fdd04b3 100644 --- a/src/openai/types/chat/completion_create_params.py +++ b/src/openai/types/chat/completion_create_params.py @@ -192,6 +192,9 @@ class CompletionCreateParamsBase(TypedDict, total=False): supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response. + + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. """ response_format: ResponseFormat diff --git a/src/openai/types/evals/create_eval_completions_run_data_source.py b/src/openai/types/evals/create_eval_completions_run_data_source.py index 74323a735e..a9f2fd0858 100644 --- a/src/openai/types/evals/create_eval_completions_run_data_source.py +++ b/src/openai/types/evals/create_eval_completions_run_data_source.py @@ -175,6 +175,9 @@ class SamplingParams(BaseModel): supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response. + + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. """ response_format: Optional[SamplingParamsResponseFormat] = None diff --git a/src/openai/types/evals/create_eval_completions_run_data_source_param.py b/src/openai/types/evals/create_eval_completions_run_data_source_param.py index 4e9c1fdeb8..e682e2db5e 100644 --- a/src/openai/types/evals/create_eval_completions_run_data_source_param.py +++ b/src/openai/types/evals/create_eval_completions_run_data_source_param.py @@ -171,6 +171,9 @@ class SamplingParams(TypedDict, total=False): supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response. + + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. """ response_format: SamplingParamsResponseFormat diff --git a/src/openai/types/evals/run_cancel_response.py b/src/openai/types/evals/run_cancel_response.py index d04d4ff657..084dd6ce5c 100644 --- a/src/openai/types/evals/run_cancel_response.py +++ b/src/openai/types/evals/run_cancel_response.py @@ -106,6 +106,9 @@ class DataSourceResponsesSourceResponses(BaseModel): supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response. + + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. """ temperature: Optional[float] = None @@ -241,6 +244,9 @@ class DataSourceResponsesSamplingParams(BaseModel): supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response. + + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. """ seed: Optional[int] = None diff --git a/src/openai/types/evals/run_create_params.py b/src/openai/types/evals/run_create_params.py index 6ff897b5de..f114fae6a4 100644 --- a/src/openai/types/evals/run_create_params.py +++ b/src/openai/types/evals/run_create_params.py @@ -119,6 +119,9 @@ class DataSourceCreateEvalResponsesRunDataSourceSourceResponses(TypedDict, total supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response. + + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. """ temperature: Optional[float] @@ -259,6 +262,9 @@ class DataSourceCreateEvalResponsesRunDataSourceSamplingParams(TypedDict, total= supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response. + + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. """ seed: int diff --git a/src/openai/types/evals/run_create_response.py b/src/openai/types/evals/run_create_response.py index defa275c8c..1343335e0d 100644 --- a/src/openai/types/evals/run_create_response.py +++ b/src/openai/types/evals/run_create_response.py @@ -106,6 +106,9 @@ class DataSourceResponsesSourceResponses(BaseModel): supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response. + + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. """ temperature: Optional[float] = None @@ -241,6 +244,9 @@ class DataSourceResponsesSamplingParams(BaseModel): supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response. + + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. """ seed: Optional[int] = None diff --git a/src/openai/types/evals/run_list_response.py b/src/openai/types/evals/run_list_response.py index 7fe0e55ace..7c32ce54a2 100644 --- a/src/openai/types/evals/run_list_response.py +++ b/src/openai/types/evals/run_list_response.py @@ -106,6 +106,9 @@ class DataSourceResponsesSourceResponses(BaseModel): supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response. + + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. """ temperature: Optional[float] = None @@ -241,6 +244,9 @@ class DataSourceResponsesSamplingParams(BaseModel): supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response. + + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. """ seed: Optional[int] = None diff --git a/src/openai/types/evals/run_retrieve_response.py b/src/openai/types/evals/run_retrieve_response.py index a50520f17d..f1212c1671 100644 --- a/src/openai/types/evals/run_retrieve_response.py +++ b/src/openai/types/evals/run_retrieve_response.py @@ -106,6 +106,9 @@ class DataSourceResponsesSourceResponses(BaseModel): supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response. + + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. """ temperature: Optional[float] = None @@ -241,6 +244,9 @@ class DataSourceResponsesSamplingParams(BaseModel): supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response. + + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. """ seed: Optional[int] = None diff --git a/src/openai/types/graders/score_model_grader.py b/src/openai/types/graders/score_model_grader.py index 908c6f91d3..35e2dc1468 100644 --- a/src/openai/types/graders/score_model_grader.py +++ b/src/openai/types/graders/score_model_grader.py @@ -70,6 +70,9 @@ class SamplingParams(BaseModel): supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response. + + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. """ seed: Optional[int] = None diff --git a/src/openai/types/graders/score_model_grader_param.py b/src/openai/types/graders/score_model_grader_param.py index 743944e099..168feeae13 100644 --- a/src/openai/types/graders/score_model_grader_param.py +++ b/src/openai/types/graders/score_model_grader_param.py @@ -76,6 +76,9 @@ class SamplingParams(TypedDict, total=False): supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response. + + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. """ seed: Optional[int] diff --git a/src/openai/types/shared/comparison_filter.py b/src/openai/types/shared/comparison_filter.py index 2ec2651ff2..33415ca4f9 100644 --- a/src/openai/types/shared/comparison_filter.py +++ b/src/openai/types/shared/comparison_filter.py @@ -1,6 +1,6 @@ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. -from typing import Union +from typing import List, Union from typing_extensions import Literal from ..._models import BaseModel @@ -13,7 +13,9 @@ class ComparisonFilter(BaseModel): """The key to compare against the value.""" type: Literal["eq", "ne", "gt", "gte", "lt", "lte"] - """Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`. + """ + Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`, `in`, + `nin`. - `eq`: equals - `ne`: not equal @@ -21,9 +23,11 @@ class ComparisonFilter(BaseModel): - `gte`: greater than or equal - `lt`: less than - `lte`: less than or equal + - `in`: in + - `nin`: not in """ - value: Union[str, float, bool] + value: Union[str, float, bool, List[Union[str, float]]] """ The value to compare against the attribute key; supports string, number, or boolean types. diff --git a/src/openai/types/shared/reasoning.py b/src/openai/types/shared/reasoning.py index 24ce301526..2cf34eb9ae 100644 --- a/src/openai/types/shared/reasoning.py +++ b/src/openai/types/shared/reasoning.py @@ -17,6 +17,9 @@ class Reasoning(BaseModel): supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response. + + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. """ generate_summary: Optional[Literal["auto", "concise", "detailed"]] = None diff --git a/src/openai/types/shared_params/comparison_filter.py b/src/openai/types/shared_params/comparison_filter.py index 38edd315ed..1c40729c19 100644 --- a/src/openai/types/shared_params/comparison_filter.py +++ b/src/openai/types/shared_params/comparison_filter.py @@ -5,6 +5,8 @@ from typing import Union from typing_extensions import Literal, Required, TypedDict +from ..._types import SequenceNotStr + __all__ = ["ComparisonFilter"] @@ -13,7 +15,9 @@ class ComparisonFilter(TypedDict, total=False): """The key to compare against the value.""" type: Required[Literal["eq", "ne", "gt", "gte", "lt", "lte"]] - """Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`. + """ + Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`, `in`, + `nin`. - `eq`: equals - `ne`: not equal @@ -21,9 +25,11 @@ class ComparisonFilter(TypedDict, total=False): - `gte`: greater than or equal - `lt`: less than - `lte`: less than or equal + - `in`: in + - `nin`: not in """ - value: Required[Union[str, float, bool]] + value: Required[Union[str, float, bool, SequenceNotStr[Union[str, float]]]] """ The value to compare against the attribute key; supports string, number, or boolean types. diff --git a/src/openai/types/shared_params/reasoning.py b/src/openai/types/shared_params/reasoning.py index 7eab2c76f7..d5461a4eaa 100644 --- a/src/openai/types/shared_params/reasoning.py +++ b/src/openai/types/shared_params/reasoning.py @@ -18,6 +18,9 @@ class Reasoning(TypedDict, total=False): supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response. + + Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + effort. """ generate_summary: Optional[Literal["auto", "concise", "detailed"]] diff --git a/src/openai/types/vector_stores/vector_store_file.py b/src/openai/types/vector_stores/vector_store_file.py index b59a61dfb0..001584dfd7 100644 --- a/src/openai/types/vector_stores/vector_store_file.py +++ b/src/openai/types/vector_stores/vector_store_file.py @@ -11,7 +11,7 @@ class LastError(BaseModel): code: Literal["server_error", "unsupported_file", "invalid_file"] - """One of `server_error` or `rate_limit_exceeded`.""" + """One of `server_error`, `unsupported_file`, or `invalid_file`.""" message: str """A human-readable description of the error.""" From 6e33c1ebc7430a1207a12439bb13738220e685ce Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Fri, 10 Oct 2025 01:10:06 +0000 Subject: [PATCH 3/3] release: 2.3.0 --- .release-please-manifest.json | 2 +- CHANGELOG.md | 13 +++++++++++++ pyproject.toml | 2 +- src/openai/_version.py | 2 +- 4 files changed, 16 insertions(+), 3 deletions(-) diff --git a/.release-please-manifest.json b/.release-please-manifest.json index bfc26f9c4e..75ec52fc91 100644 --- a/.release-please-manifest.json +++ b/.release-please-manifest.json @@ -1,3 +1,3 @@ { - ".": "2.2.0" + ".": "2.3.0" } \ No newline at end of file diff --git a/CHANGELOG.md b/CHANGELOG.md index 336c6601a2..66eec00fea 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,18 @@ # Changelog +## 2.3.0 (2025-10-10) + +Full Changelog: [v2.2.0...v2.3.0](https://github.com/openai/openai-python/compare/v2.2.0...v2.3.0) + +### Features + +* **api:** comparison filter in/not in ([aa49f62](https://github.com/openai/openai-python/commit/aa49f626a6ea9d77ad008badfb3741e16232d62f)) + + +### Chores + +* **package:** bump jiter to >=0.10.0 to support Python 3.14 ([#2618](https://github.com/openai/openai-python/issues/2618)) ([aa445ca](https://github.com/openai/openai-python/commit/aa445cab5c93c6908697fe98e73e16963330b141)) + ## 2.2.0 (2025-10-06) Full Changelog: [v2.1.0...v2.2.0](https://github.com/openai/openai-python/compare/v2.1.0...v2.2.0) diff --git a/pyproject.toml b/pyproject.toml index bb5eb7cf41..7197505229 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "openai" -version = "2.2.0" +version = "2.3.0" description = "The official Python library for the openai API" dynamic = ["readme"] license = "Apache-2.0" diff --git a/src/openai/_version.py b/src/openai/_version.py index 2fdb2912ef..f202a6d61c 100644 --- a/src/openai/_version.py +++ b/src/openai/_version.py @@ -1,4 +1,4 @@ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. __title__ = "openai" -__version__ = "2.2.0" # x-release-please-version +__version__ = "2.3.0" # x-release-please-version