@@ -77,6 +77,7 @@ def create(
77
77
* ,
78
78
input : Union [str , ResponseInputParam ],
79
79
model : ResponsesModel ,
80
+ background : Optional [bool ] | NotGiven = NOT_GIVEN ,
80
81
include : Optional [List [ResponseIncludable ]] | NotGiven = NOT_GIVEN ,
81
82
instructions : Optional [str ] | NotGiven = NOT_GIVEN ,
82
83
max_output_tokens : Optional [int ] | NotGiven = NOT_GIVEN ,
@@ -132,6 +133,9 @@ def create(
132
133
[model guide](https://platform.openai.com/docs/models) to browse and compare
133
134
available models.
134
135
136
+ background: Whether to run the model response in the background.
137
+ [Learn more](https://platform.openai.com/docs/guides/background).
138
+
135
139
include: Specify additional output data to include in the model response. Currently
136
140
supported values are:
137
141
@@ -267,6 +271,7 @@ def create(
267
271
input : Union [str , ResponseInputParam ],
268
272
model : ResponsesModel ,
269
273
stream : Literal [True ],
274
+ background : Optional [bool ] | NotGiven = NOT_GIVEN ,
270
275
include : Optional [List [ResponseIncludable ]] | NotGiven = NOT_GIVEN ,
271
276
instructions : Optional [str ] | NotGiven = NOT_GIVEN ,
272
277
max_output_tokens : Optional [int ] | NotGiven = NOT_GIVEN ,
@@ -328,6 +333,9 @@ def create(
328
333
[Streaming section below](https://platform.openai.com/docs/api-reference/responses-streaming)
329
334
for more information.
330
335
336
+ background: Whether to run the model response in the background.
337
+ [Learn more](https://platform.openai.com/docs/guides/background).
338
+
331
339
include: Specify additional output data to include in the model response. Currently
332
340
supported values are:
333
341
@@ -456,6 +464,7 @@ def create(
456
464
input : Union [str , ResponseInputParam ],
457
465
model : ResponsesModel ,
458
466
stream : bool ,
467
+ background : Optional [bool ] | NotGiven = NOT_GIVEN ,
459
468
include : Optional [List [ResponseIncludable ]] | NotGiven = NOT_GIVEN ,
460
469
instructions : Optional [str ] | NotGiven = NOT_GIVEN ,
461
470
max_output_tokens : Optional [int ] | NotGiven = NOT_GIVEN ,
@@ -517,6 +526,9 @@ def create(
517
526
[Streaming section below](https://platform.openai.com/docs/api-reference/responses-streaming)
518
527
for more information.
519
528
529
+ background: Whether to run the model response in the background.
530
+ [Learn more](https://platform.openai.com/docs/guides/background).
531
+
520
532
include: Specify additional output data to include in the model response. Currently
521
533
supported values are:
522
534
@@ -644,6 +656,7 @@ def create(
644
656
* ,
645
657
input : Union [str , ResponseInputParam ],
646
658
model : ResponsesModel ,
659
+ background : Optional [bool ] | NotGiven = NOT_GIVEN ,
647
660
include : Optional [List [ResponseIncludable ]] | NotGiven = NOT_GIVEN ,
648
661
instructions : Optional [str ] | NotGiven = NOT_GIVEN ,
649
662
max_output_tokens : Optional [int ] | NotGiven = NOT_GIVEN ,
@@ -674,6 +687,7 @@ def create(
674
687
{
675
688
"input" : input ,
676
689
"model" : model ,
690
+ "background" : background ,
677
691
"include" : include ,
678
692
"instructions" : instructions ,
679
693
"max_output_tokens" : max_output_tokens ,
@@ -965,6 +979,7 @@ async def create(
965
979
* ,
966
980
input : Union [str , ResponseInputParam ],
967
981
model : ResponsesModel ,
982
+ background : Optional [bool ] | NotGiven = NOT_GIVEN ,
968
983
include : Optional [List [ResponseIncludable ]] | NotGiven = NOT_GIVEN ,
969
984
instructions : Optional [str ] | NotGiven = NOT_GIVEN ,
970
985
max_output_tokens : Optional [int ] | NotGiven = NOT_GIVEN ,
@@ -1020,6 +1035,9 @@ async def create(
1020
1035
[model guide](https://platform.openai.com/docs/models) to browse and compare
1021
1036
available models.
1022
1037
1038
+ background: Whether to run the model response in the background.
1039
+ [Learn more](https://platform.openai.com/docs/guides/background).
1040
+
1023
1041
include: Specify additional output data to include in the model response. Currently
1024
1042
supported values are:
1025
1043
@@ -1155,6 +1173,7 @@ async def create(
1155
1173
input : Union [str , ResponseInputParam ],
1156
1174
model : ResponsesModel ,
1157
1175
stream : Literal [True ],
1176
+ background : Optional [bool ] | NotGiven = NOT_GIVEN ,
1158
1177
include : Optional [List [ResponseIncludable ]] | NotGiven = NOT_GIVEN ,
1159
1178
instructions : Optional [str ] | NotGiven = NOT_GIVEN ,
1160
1179
max_output_tokens : Optional [int ] | NotGiven = NOT_GIVEN ,
@@ -1216,6 +1235,9 @@ async def create(
1216
1235
[Streaming section below](https://platform.openai.com/docs/api-reference/responses-streaming)
1217
1236
for more information.
1218
1237
1238
+ background: Whether to run the model response in the background.
1239
+ [Learn more](https://platform.openai.com/docs/guides/background).
1240
+
1219
1241
include: Specify additional output data to include in the model response. Currently
1220
1242
supported values are:
1221
1243
@@ -1344,6 +1366,7 @@ async def create(
1344
1366
input : Union [str , ResponseInputParam ],
1345
1367
model : ResponsesModel ,
1346
1368
stream : bool ,
1369
+ background : Optional [bool ] | NotGiven = NOT_GIVEN ,
1347
1370
include : Optional [List [ResponseIncludable ]] | NotGiven = NOT_GIVEN ,
1348
1371
instructions : Optional [str ] | NotGiven = NOT_GIVEN ,
1349
1372
max_output_tokens : Optional [int ] | NotGiven = NOT_GIVEN ,
@@ -1405,6 +1428,9 @@ async def create(
1405
1428
[Streaming section below](https://platform.openai.com/docs/api-reference/responses-streaming)
1406
1429
for more information.
1407
1430
1431
+ background: Whether to run the model response in the background.
1432
+ [Learn more](https://platform.openai.com/docs/guides/background).
1433
+
1408
1434
include: Specify additional output data to include in the model response. Currently
1409
1435
supported values are:
1410
1436
@@ -1532,6 +1558,7 @@ async def create(
1532
1558
* ,
1533
1559
input : Union [str , ResponseInputParam ],
1534
1560
model : ResponsesModel ,
1561
+ background : Optional [bool ] | NotGiven = NOT_GIVEN ,
1535
1562
include : Optional [List [ResponseIncludable ]] | NotGiven = NOT_GIVEN ,
1536
1563
instructions : Optional [str ] | NotGiven = NOT_GIVEN ,
1537
1564
max_output_tokens : Optional [int ] | NotGiven = NOT_GIVEN ,
@@ -1562,6 +1589,7 @@ async def create(
1562
1589
{
1563
1590
"input" : input ,
1564
1591
"model" : model ,
1592
+ "background" : background ,
1565
1593
"include" : include ,
1566
1594
"instructions" : instructions ,
1567
1595
"max_output_tokens" : max_output_tokens ,
0 commit comments