Skip to content

Commit dc97d05

Browse files
stainless-app[bot]stainless-bot
authored andcommitted
fix(api): add missing reasoning effort + model enums (#2096)
1 parent 0288213 commit dc97d05

File tree

11 files changed

+248
-26
lines changed

11 files changed

+248
-26
lines changed

.stats.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,2 +1,2 @@
11
configured_endpoints: 69
2-
openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/openai-fc5dbc19505b0035f9e7f88868619f4fb519b048bde011f6154f3132d4be71fb.yml
2+
openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/openai-7c699d4503077d06a4a44f52c0c1f902d19a87c766b8be75b97c8dfd484ad4aa.yml

src/openai/resources/beta/assistants.py

Lines changed: 104 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -61,6 +61,7 @@ def create(
6161
instructions: Optional[str] | NotGiven = NOT_GIVEN,
6262
metadata: Optional[Metadata] | NotGiven = NOT_GIVEN,
6363
name: Optional[str] | NotGiven = NOT_GIVEN,
64+
reasoning_effort: Optional[Literal["low", "medium", "high"]] | NotGiven = NOT_GIVEN,
6465
response_format: Optional[AssistantResponseFormatOptionParam] | NotGiven = NOT_GIVEN,
6566
temperature: Optional[float] | NotGiven = NOT_GIVEN,
6667
tool_resources: Optional[assistant_create_params.ToolResources] | NotGiven = NOT_GIVEN,
@@ -97,6 +98,13 @@ def create(
9798
9899
name: The name of the assistant. The maximum length is 256 characters.
99100
101+
reasoning_effort: **o1 and o3-mini models only**
102+
103+
Constrains effort on reasoning for
104+
[reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
105+
supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
106+
result in faster responses and fewer tokens used on reasoning in a response.
107+
100108
response_format: Specifies the format that the model must output. Compatible with
101109
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
102110
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -155,6 +163,7 @@ def create(
155163
"instructions": instructions,
156164
"metadata": metadata,
157165
"name": name,
166+
"reasoning_effort": reasoning_effort,
158167
"response_format": response_format,
159168
"temperature": temperature,
160169
"tool_resources": tool_resources,
@@ -210,8 +219,42 @@ def update(
210219
description: Optional[str] | NotGiven = NOT_GIVEN,
211220
instructions: Optional[str] | NotGiven = NOT_GIVEN,
212221
metadata: Optional[Metadata] | NotGiven = NOT_GIVEN,
213-
model: str | NotGiven = NOT_GIVEN,
222+
model: Union[
223+
str,
224+
Literal[
225+
"o3-mini",
226+
"o3-mini-2025-01-31",
227+
"o1",
228+
"o1-2024-12-17",
229+
"gpt-4o",
230+
"gpt-4o-2024-11-20",
231+
"gpt-4o-2024-08-06",
232+
"gpt-4o-2024-05-13",
233+
"gpt-4o-mini",
234+
"gpt-4o-mini-2024-07-18",
235+
"gpt-4-turbo",
236+
"gpt-4-turbo-2024-04-09",
237+
"gpt-4-0125-preview",
238+
"gpt-4-turbo-preview",
239+
"gpt-4-1106-preview",
240+
"gpt-4-vision-preview",
241+
"gpt-4",
242+
"gpt-4-0314",
243+
"gpt-4-0613",
244+
"gpt-4-32k",
245+
"gpt-4-32k-0314",
246+
"gpt-4-32k-0613",
247+
"gpt-3.5-turbo",
248+
"gpt-3.5-turbo-16k",
249+
"gpt-3.5-turbo-0613",
250+
"gpt-3.5-turbo-1106",
251+
"gpt-3.5-turbo-0125",
252+
"gpt-3.5-turbo-16k-0613",
253+
],
254+
]
255+
| NotGiven = NOT_GIVEN,
214256
name: Optional[str] | NotGiven = NOT_GIVEN,
257+
reasoning_effort: Optional[Literal["low", "medium", "high"]] | NotGiven = NOT_GIVEN,
215258
response_format: Optional[AssistantResponseFormatOptionParam] | NotGiven = NOT_GIVEN,
216259
temperature: Optional[float] | NotGiven = NOT_GIVEN,
217260
tool_resources: Optional[assistant_update_params.ToolResources] | NotGiven = NOT_GIVEN,
@@ -249,6 +292,13 @@ def update(
249292
250293
name: The name of the assistant. The maximum length is 256 characters.
251294
295+
reasoning_effort: **o1 and o3-mini models only**
296+
297+
Constrains effort on reasoning for
298+
[reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
299+
supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
300+
result in faster responses and fewer tokens used on reasoning in a response.
301+
252302
response_format: Specifies the format that the model must output. Compatible with
253303
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
254304
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -309,6 +359,7 @@ def update(
309359
"metadata": metadata,
310360
"model": model,
311361
"name": name,
362+
"reasoning_effort": reasoning_effort,
312363
"response_format": response_format,
313364
"temperature": temperature,
314365
"tool_resources": tool_resources,
@@ -451,6 +502,7 @@ async def create(
451502
instructions: Optional[str] | NotGiven = NOT_GIVEN,
452503
metadata: Optional[Metadata] | NotGiven = NOT_GIVEN,
453504
name: Optional[str] | NotGiven = NOT_GIVEN,
505+
reasoning_effort: Optional[Literal["low", "medium", "high"]] | NotGiven = NOT_GIVEN,
454506
response_format: Optional[AssistantResponseFormatOptionParam] | NotGiven = NOT_GIVEN,
455507
temperature: Optional[float] | NotGiven = NOT_GIVEN,
456508
tool_resources: Optional[assistant_create_params.ToolResources] | NotGiven = NOT_GIVEN,
@@ -487,6 +539,13 @@ async def create(
487539
488540
name: The name of the assistant. The maximum length is 256 characters.
489541
542+
reasoning_effort: **o1 and o3-mini models only**
543+
544+
Constrains effort on reasoning for
545+
[reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
546+
supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
547+
result in faster responses and fewer tokens used on reasoning in a response.
548+
490549
response_format: Specifies the format that the model must output. Compatible with
491550
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
492551
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -545,6 +604,7 @@ async def create(
545604
"instructions": instructions,
546605
"metadata": metadata,
547606
"name": name,
607+
"reasoning_effort": reasoning_effort,
548608
"response_format": response_format,
549609
"temperature": temperature,
550610
"tool_resources": tool_resources,
@@ -600,8 +660,42 @@ async def update(
600660
description: Optional[str] | NotGiven = NOT_GIVEN,
601661
instructions: Optional[str] | NotGiven = NOT_GIVEN,
602662
metadata: Optional[Metadata] | NotGiven = NOT_GIVEN,
603-
model: str | NotGiven = NOT_GIVEN,
663+
model: Union[
664+
str,
665+
Literal[
666+
"o3-mini",
667+
"o3-mini-2025-01-31",
668+
"o1",
669+
"o1-2024-12-17",
670+
"gpt-4o",
671+
"gpt-4o-2024-11-20",
672+
"gpt-4o-2024-08-06",
673+
"gpt-4o-2024-05-13",
674+
"gpt-4o-mini",
675+
"gpt-4o-mini-2024-07-18",
676+
"gpt-4-turbo",
677+
"gpt-4-turbo-2024-04-09",
678+
"gpt-4-0125-preview",
679+
"gpt-4-turbo-preview",
680+
"gpt-4-1106-preview",
681+
"gpt-4-vision-preview",
682+
"gpt-4",
683+
"gpt-4-0314",
684+
"gpt-4-0613",
685+
"gpt-4-32k",
686+
"gpt-4-32k-0314",
687+
"gpt-4-32k-0613",
688+
"gpt-3.5-turbo",
689+
"gpt-3.5-turbo-16k",
690+
"gpt-3.5-turbo-0613",
691+
"gpt-3.5-turbo-1106",
692+
"gpt-3.5-turbo-0125",
693+
"gpt-3.5-turbo-16k-0613",
694+
],
695+
]
696+
| NotGiven = NOT_GIVEN,
604697
name: Optional[str] | NotGiven = NOT_GIVEN,
698+
reasoning_effort: Optional[Literal["low", "medium", "high"]] | NotGiven = NOT_GIVEN,
605699
response_format: Optional[AssistantResponseFormatOptionParam] | NotGiven = NOT_GIVEN,
606700
temperature: Optional[float] | NotGiven = NOT_GIVEN,
607701
tool_resources: Optional[assistant_update_params.ToolResources] | NotGiven = NOT_GIVEN,
@@ -639,6 +733,13 @@ async def update(
639733
640734
name: The name of the assistant. The maximum length is 256 characters.
641735
736+
reasoning_effort: **o1 and o3-mini models only**
737+
738+
Constrains effort on reasoning for
739+
[reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
740+
supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
741+
result in faster responses and fewer tokens used on reasoning in a response.
742+
642743
response_format: Specifies the format that the model must output. Compatible with
643744
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
644745
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -699,6 +800,7 @@ async def update(
699800
"metadata": metadata,
700801
"model": model,
701802
"name": name,
803+
"reasoning_effort": reasoning_effort,
702804
"response_format": response_format,
703805
"temperature": temperature,
704806
"tool_resources": tool_resources,

src/openai/resources/beta/threads/runs/runs.py

Lines changed: 52 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -96,6 +96,7 @@ def create(
9696
metadata: Optional[Metadata] | NotGiven = NOT_GIVEN,
9797
model: Union[str, ChatModel, None] | NotGiven = NOT_GIVEN,
9898
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
99+
reasoning_effort: Optional[Literal["low", "medium", "high"]] | NotGiven = NOT_GIVEN,
99100
response_format: Optional[AssistantResponseFormatOptionParam] | NotGiven = NOT_GIVEN,
100101
stream: Optional[Literal[False]] | NotGiven = NOT_GIVEN,
101102
temperature: Optional[float] | NotGiven = NOT_GIVEN,
@@ -164,6 +165,13 @@ def create(
164165
[parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
165166
during tool use.
166167
168+
reasoning_effort: **o1 and o3-mini models only**
169+
170+
Constrains effort on reasoning for
171+
[reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
172+
supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
173+
result in faster responses and fewer tokens used on reasoning in a response.
174+
167175
response_format: Specifies the format that the model must output. Compatible with
168176
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
169177
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -239,6 +247,7 @@ def create(
239247
metadata: Optional[Metadata] | NotGiven = NOT_GIVEN,
240248
model: Union[str, ChatModel, None] | NotGiven = NOT_GIVEN,
241249
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
250+
reasoning_effort: Optional[Literal["low", "medium", "high"]] | NotGiven = NOT_GIVEN,
242251
response_format: Optional[AssistantResponseFormatOptionParam] | NotGiven = NOT_GIVEN,
243252
temperature: Optional[float] | NotGiven = NOT_GIVEN,
244253
tool_choice: Optional[AssistantToolChoiceOptionParam] | NotGiven = NOT_GIVEN,
@@ -310,6 +319,13 @@ def create(
310319
[parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
311320
during tool use.
312321
322+
reasoning_effort: **o1 and o3-mini models only**
323+
324+
Constrains effort on reasoning for
325+
[reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
326+
supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
327+
result in faster responses and fewer tokens used on reasoning in a response.
328+
313329
response_format: Specifies the format that the model must output. Compatible with
314330
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
315331
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -381,6 +397,7 @@ def create(
381397
metadata: Optional[Metadata] | NotGiven = NOT_GIVEN,
382398
model: Union[str, ChatModel, None] | NotGiven = NOT_GIVEN,
383399
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
400+
reasoning_effort: Optional[Literal["low", "medium", "high"]] | NotGiven = NOT_GIVEN,
384401
response_format: Optional[AssistantResponseFormatOptionParam] | NotGiven = NOT_GIVEN,
385402
temperature: Optional[float] | NotGiven = NOT_GIVEN,
386403
tool_choice: Optional[AssistantToolChoiceOptionParam] | NotGiven = NOT_GIVEN,
@@ -452,6 +469,13 @@ def create(
452469
[parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
453470
during tool use.
454471
472+
reasoning_effort: **o1 and o3-mini models only**
473+
474+
Constrains effort on reasoning for
475+
[reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
476+
supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
477+
result in faster responses and fewer tokens used on reasoning in a response.
478+
455479
response_format: Specifies the format that the model must output. Compatible with
456480
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
457481
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -522,6 +546,7 @@ def create(
522546
metadata: Optional[Metadata] | NotGiven = NOT_GIVEN,
523547
model: Union[str, ChatModel, None] | NotGiven = NOT_GIVEN,
524548
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
549+
reasoning_effort: Optional[Literal["low", "medium", "high"]] | NotGiven = NOT_GIVEN,
525550
response_format: Optional[AssistantResponseFormatOptionParam] | NotGiven = NOT_GIVEN,
526551
stream: Optional[Literal[False]] | Literal[True] | NotGiven = NOT_GIVEN,
527552
temperature: Optional[float] | NotGiven = NOT_GIVEN,
@@ -552,6 +577,7 @@ def create(
552577
"metadata": metadata,
553578
"model": model,
554579
"parallel_tool_calls": parallel_tool_calls,
580+
"reasoning_effort": reasoning_effort,
555581
"response_format": response_format,
556582
"stream": stream,
557583
"temperature": temperature,
@@ -1469,6 +1495,7 @@ async def create(
14691495
metadata: Optional[Metadata] | NotGiven = NOT_GIVEN,
14701496
model: Union[str, ChatModel, None] | NotGiven = NOT_GIVEN,
14711497
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
1498+
reasoning_effort: Optional[Literal["low", "medium", "high"]] | NotGiven = NOT_GIVEN,
14721499
response_format: Optional[AssistantResponseFormatOptionParam] | NotGiven = NOT_GIVEN,
14731500
stream: Optional[Literal[False]] | NotGiven = NOT_GIVEN,
14741501
temperature: Optional[float] | NotGiven = NOT_GIVEN,
@@ -1537,6 +1564,13 @@ async def create(
15371564
[parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
15381565
during tool use.
15391566
1567+
reasoning_effort: **o1 and o3-mini models only**
1568+
1569+
Constrains effort on reasoning for
1570+
[reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
1571+
supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
1572+
result in faster responses and fewer tokens used on reasoning in a response.
1573+
15401574
response_format: Specifies the format that the model must output. Compatible with
15411575
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
15421576
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -1612,6 +1646,7 @@ async def create(
16121646
metadata: Optional[Metadata] | NotGiven = NOT_GIVEN,
16131647
model: Union[str, ChatModel, None] | NotGiven = NOT_GIVEN,
16141648
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
1649+
reasoning_effort: Optional[Literal["low", "medium", "high"]] | NotGiven = NOT_GIVEN,
16151650
response_format: Optional[AssistantResponseFormatOptionParam] | NotGiven = NOT_GIVEN,
16161651
temperature: Optional[float] | NotGiven = NOT_GIVEN,
16171652
tool_choice: Optional[AssistantToolChoiceOptionParam] | NotGiven = NOT_GIVEN,
@@ -1683,6 +1718,13 @@ async def create(
16831718
[parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
16841719
during tool use.
16851720
1721+
reasoning_effort: **o1 and o3-mini models only**
1722+
1723+
Constrains effort on reasoning for
1724+
[reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
1725+
supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
1726+
result in faster responses and fewer tokens used on reasoning in a response.
1727+
16861728
response_format: Specifies the format that the model must output. Compatible with
16871729
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
16881730
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -1754,6 +1796,7 @@ async def create(
17541796
metadata: Optional[Metadata] | NotGiven = NOT_GIVEN,
17551797
model: Union[str, ChatModel, None] | NotGiven = NOT_GIVEN,
17561798
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
1799+
reasoning_effort: Optional[Literal["low", "medium", "high"]] | NotGiven = NOT_GIVEN,
17571800
response_format: Optional[AssistantResponseFormatOptionParam] | NotGiven = NOT_GIVEN,
17581801
temperature: Optional[float] | NotGiven = NOT_GIVEN,
17591802
tool_choice: Optional[AssistantToolChoiceOptionParam] | NotGiven = NOT_GIVEN,
@@ -1825,6 +1868,13 @@ async def create(
18251868
[parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
18261869
during tool use.
18271870
1871+
reasoning_effort: **o1 and o3-mini models only**
1872+
1873+
Constrains effort on reasoning for
1874+
[reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
1875+
supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
1876+
result in faster responses and fewer tokens used on reasoning in a response.
1877+
18281878
response_format: Specifies the format that the model must output. Compatible with
18291879
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
18301880
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -1895,6 +1945,7 @@ async def create(
18951945
metadata: Optional[Metadata] | NotGiven = NOT_GIVEN,
18961946
model: Union[str, ChatModel, None] | NotGiven = NOT_GIVEN,
18971947
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
1948+
reasoning_effort: Optional[Literal["low", "medium", "high"]] | NotGiven = NOT_GIVEN,
18981949
response_format: Optional[AssistantResponseFormatOptionParam] | NotGiven = NOT_GIVEN,
18991950
stream: Optional[Literal[False]] | Literal[True] | NotGiven = NOT_GIVEN,
19001951
temperature: Optional[float] | NotGiven = NOT_GIVEN,
@@ -1925,6 +1976,7 @@ async def create(
19251976
"metadata": metadata,
19261977
"model": model,
19271978
"parallel_tool_calls": parallel_tool_calls,
1979+
"reasoning_effort": reasoning_effort,
19281980
"response_format": response_format,
19291981
"stream": stream,
19301982
"temperature": temperature,

0 commit comments

Comments
 (0)