@@ -96,6 +96,7 @@ def create(
96
96
metadata : Optional [Metadata ] | NotGiven = NOT_GIVEN ,
97
97
model : Union [str , ChatModel , None ] | NotGiven = NOT_GIVEN ,
98
98
parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
99
+ reasoning_effort : Optional [Literal ["low" , "medium" , "high" ]] | NotGiven = NOT_GIVEN ,
99
100
response_format : Optional [AssistantResponseFormatOptionParam ] | NotGiven = NOT_GIVEN ,
100
101
stream : Optional [Literal [False ]] | NotGiven = NOT_GIVEN ,
101
102
temperature : Optional [float ] | NotGiven = NOT_GIVEN ,
@@ -164,6 +165,13 @@ def create(
164
165
[parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
165
166
during tool use.
166
167
168
+ reasoning_effort: **o1 and o3-mini models only**
169
+
170
+ Constrains effort on reasoning for
171
+ [reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
172
+ supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
173
+ result in faster responses and fewer tokens used on reasoning in a response.
174
+
167
175
response_format: Specifies the format that the model must output. Compatible with
168
176
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
169
177
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -239,6 +247,7 @@ def create(
239
247
metadata : Optional [Metadata ] | NotGiven = NOT_GIVEN ,
240
248
model : Union [str , ChatModel , None ] | NotGiven = NOT_GIVEN ,
241
249
parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
250
+ reasoning_effort : Optional [Literal ["low" , "medium" , "high" ]] | NotGiven = NOT_GIVEN ,
242
251
response_format : Optional [AssistantResponseFormatOptionParam ] | NotGiven = NOT_GIVEN ,
243
252
temperature : Optional [float ] | NotGiven = NOT_GIVEN ,
244
253
tool_choice : Optional [AssistantToolChoiceOptionParam ] | NotGiven = NOT_GIVEN ,
@@ -310,6 +319,13 @@ def create(
310
319
[parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
311
320
during tool use.
312
321
322
+ reasoning_effort: **o1 and o3-mini models only**
323
+
324
+ Constrains effort on reasoning for
325
+ [reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
326
+ supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
327
+ result in faster responses and fewer tokens used on reasoning in a response.
328
+
313
329
response_format: Specifies the format that the model must output. Compatible with
314
330
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
315
331
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -381,6 +397,7 @@ def create(
381
397
metadata : Optional [Metadata ] | NotGiven = NOT_GIVEN ,
382
398
model : Union [str , ChatModel , None ] | NotGiven = NOT_GIVEN ,
383
399
parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
400
+ reasoning_effort : Optional [Literal ["low" , "medium" , "high" ]] | NotGiven = NOT_GIVEN ,
384
401
response_format : Optional [AssistantResponseFormatOptionParam ] | NotGiven = NOT_GIVEN ,
385
402
temperature : Optional [float ] | NotGiven = NOT_GIVEN ,
386
403
tool_choice : Optional [AssistantToolChoiceOptionParam ] | NotGiven = NOT_GIVEN ,
@@ -452,6 +469,13 @@ def create(
452
469
[parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
453
470
during tool use.
454
471
472
+ reasoning_effort: **o1 and o3-mini models only**
473
+
474
+ Constrains effort on reasoning for
475
+ [reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
476
+ supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
477
+ result in faster responses and fewer tokens used on reasoning in a response.
478
+
455
479
response_format: Specifies the format that the model must output. Compatible with
456
480
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
457
481
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -522,6 +546,7 @@ def create(
522
546
metadata : Optional [Metadata ] | NotGiven = NOT_GIVEN ,
523
547
model : Union [str , ChatModel , None ] | NotGiven = NOT_GIVEN ,
524
548
parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
549
+ reasoning_effort : Optional [Literal ["low" , "medium" , "high" ]] | NotGiven = NOT_GIVEN ,
525
550
response_format : Optional [AssistantResponseFormatOptionParam ] | NotGiven = NOT_GIVEN ,
526
551
stream : Optional [Literal [False ]] | Literal [True ] | NotGiven = NOT_GIVEN ,
527
552
temperature : Optional [float ] | NotGiven = NOT_GIVEN ,
@@ -552,6 +577,7 @@ def create(
552
577
"metadata" : metadata ,
553
578
"model" : model ,
554
579
"parallel_tool_calls" : parallel_tool_calls ,
580
+ "reasoning_effort" : reasoning_effort ,
555
581
"response_format" : response_format ,
556
582
"stream" : stream ,
557
583
"temperature" : temperature ,
@@ -1469,6 +1495,7 @@ async def create(
1469
1495
metadata : Optional [Metadata ] | NotGiven = NOT_GIVEN ,
1470
1496
model : Union [str , ChatModel , None ] | NotGiven = NOT_GIVEN ,
1471
1497
parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
1498
+ reasoning_effort : Optional [Literal ["low" , "medium" , "high" ]] | NotGiven = NOT_GIVEN ,
1472
1499
response_format : Optional [AssistantResponseFormatOptionParam ] | NotGiven = NOT_GIVEN ,
1473
1500
stream : Optional [Literal [False ]] | NotGiven = NOT_GIVEN ,
1474
1501
temperature : Optional [float ] | NotGiven = NOT_GIVEN ,
@@ -1537,6 +1564,13 @@ async def create(
1537
1564
[parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
1538
1565
during tool use.
1539
1566
1567
+ reasoning_effort: **o1 and o3-mini models only**
1568
+
1569
+ Constrains effort on reasoning for
1570
+ [reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
1571
+ supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
1572
+ result in faster responses and fewer tokens used on reasoning in a response.
1573
+
1540
1574
response_format: Specifies the format that the model must output. Compatible with
1541
1575
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
1542
1576
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -1612,6 +1646,7 @@ async def create(
1612
1646
metadata : Optional [Metadata ] | NotGiven = NOT_GIVEN ,
1613
1647
model : Union [str , ChatModel , None ] | NotGiven = NOT_GIVEN ,
1614
1648
parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
1649
+ reasoning_effort : Optional [Literal ["low" , "medium" , "high" ]] | NotGiven = NOT_GIVEN ,
1615
1650
response_format : Optional [AssistantResponseFormatOptionParam ] | NotGiven = NOT_GIVEN ,
1616
1651
temperature : Optional [float ] | NotGiven = NOT_GIVEN ,
1617
1652
tool_choice : Optional [AssistantToolChoiceOptionParam ] | NotGiven = NOT_GIVEN ,
@@ -1683,6 +1718,13 @@ async def create(
1683
1718
[parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
1684
1719
during tool use.
1685
1720
1721
+ reasoning_effort: **o1 and o3-mini models only**
1722
+
1723
+ Constrains effort on reasoning for
1724
+ [reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
1725
+ supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
1726
+ result in faster responses and fewer tokens used on reasoning in a response.
1727
+
1686
1728
response_format: Specifies the format that the model must output. Compatible with
1687
1729
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
1688
1730
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -1754,6 +1796,7 @@ async def create(
1754
1796
metadata : Optional [Metadata ] | NotGiven = NOT_GIVEN ,
1755
1797
model : Union [str , ChatModel , None ] | NotGiven = NOT_GIVEN ,
1756
1798
parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
1799
+ reasoning_effort : Optional [Literal ["low" , "medium" , "high" ]] | NotGiven = NOT_GIVEN ,
1757
1800
response_format : Optional [AssistantResponseFormatOptionParam ] | NotGiven = NOT_GIVEN ,
1758
1801
temperature : Optional [float ] | NotGiven = NOT_GIVEN ,
1759
1802
tool_choice : Optional [AssistantToolChoiceOptionParam ] | NotGiven = NOT_GIVEN ,
@@ -1825,6 +1868,13 @@ async def create(
1825
1868
[parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
1826
1869
during tool use.
1827
1870
1871
+ reasoning_effort: **o1 and o3-mini models only**
1872
+
1873
+ Constrains effort on reasoning for
1874
+ [reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
1875
+ supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
1876
+ result in faster responses and fewer tokens used on reasoning in a response.
1877
+
1828
1878
response_format: Specifies the format that the model must output. Compatible with
1829
1879
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
1830
1880
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -1895,6 +1945,7 @@ async def create(
1895
1945
metadata : Optional [Metadata ] | NotGiven = NOT_GIVEN ,
1896
1946
model : Union [str , ChatModel , None ] | NotGiven = NOT_GIVEN ,
1897
1947
parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
1948
+ reasoning_effort : Optional [Literal ["low" , "medium" , "high" ]] | NotGiven = NOT_GIVEN ,
1898
1949
response_format : Optional [AssistantResponseFormatOptionParam ] | NotGiven = NOT_GIVEN ,
1899
1950
stream : Optional [Literal [False ]] | Literal [True ] | NotGiven = NOT_GIVEN ,
1900
1951
temperature : Optional [float ] | NotGiven = NOT_GIVEN ,
@@ -1925,6 +1976,7 @@ async def create(
1925
1976
"metadata" : metadata ,
1926
1977
"model" : model ,
1927
1978
"parallel_tool_calls" : parallel_tool_calls ,
1979
+ "reasoning_effort" : reasoning_effort ,
1928
1980
"response_format" : response_format ,
1929
1981
"stream" : stream ,
1930
1982
"temperature" : temperature ,
0 commit comments