File tree Expand file tree Collapse file tree 1 file changed +4
-0
lines changed
pydantic_ai_slim/pydantic_ai Expand file tree Collapse file tree 1 file changed +4
-0
lines changed Original file line number Diff line number Diff line change @@ -22,6 +22,7 @@ class ModelSettings(TypedDict, total=False):
22
22
"""The maximum number of tokens to generate before stopping.
23
23
24
24
Supported by:
25
+
25
26
* Gemini
26
27
* Anthropic
27
28
* OpenAI
@@ -37,6 +38,7 @@ class ModelSettings(TypedDict, total=False):
37
38
Note that even with `temperature` of `0.0`, the results will not be fully deterministic.
38
39
39
40
Supported by:
41
+
40
42
* Gemini
41
43
* Anthropic
42
44
* OpenAI
@@ -51,6 +53,7 @@ class ModelSettings(TypedDict, total=False):
51
53
You should either alter `temperature` or `top_p`, but not both.
52
54
53
55
Supported by:
56
+
54
57
* Gemini
55
58
* Anthropic
56
59
* OpenAI
@@ -61,6 +64,7 @@ class ModelSettings(TypedDict, total=False):
61
64
"""Override the client-level default timeout for a request, in seconds.
62
65
63
66
Supported by:
67
+
64
68
* Gemini
65
69
* Anthropic
66
70
* OpenAI
You can’t perform that action at this time.
0 commit comments