@@ -122,6 +122,7 @@ def test_method_create(self, client: PromptFoundry) -> None:
122
122
],
123
123
name = "string" ,
124
124
parameters = {
125
+ "model_provider" : "OPENAI" ,
125
126
"model_name" : "string" ,
126
127
"response_format" : "JSON" ,
127
128
"temperature" : 0 ,
@@ -238,6 +239,7 @@ def test_raw_response_create(self, client: PromptFoundry) -> None:
238
239
],
239
240
name = "string" ,
240
241
parameters = {
242
+ "model_provider" : "OPENAI" ,
241
243
"model_name" : "string" ,
242
244
"response_format" : "JSON" ,
243
245
"temperature" : 0 ,
@@ -358,6 +360,7 @@ def test_streaming_response_create(self, client: PromptFoundry) -> None:
358
360
],
359
361
name = "string" ,
360
362
parameters = {
363
+ "model_provider" : "OPENAI" ,
361
364
"model_name" : "string" ,
362
365
"response_format" : "JSON" ,
363
366
"temperature" : 0 ,
@@ -481,6 +484,7 @@ def test_method_update(self, client: PromptFoundry) -> None:
481
484
],
482
485
name = "string" ,
483
486
parameters = {
487
+ "model_provider" : "OPENAI" ,
484
488
"model_name" : "string" ,
485
489
"response_format" : "JSON" ,
486
490
"temperature" : 0 ,
@@ -598,6 +602,7 @@ def test_raw_response_update(self, client: PromptFoundry) -> None:
598
602
],
599
603
name = "string" ,
600
604
parameters = {
605
+ "model_provider" : "OPENAI" ,
601
606
"model_name" : "string" ,
602
607
"response_format" : "JSON" ,
603
608
"temperature" : 0 ,
@@ -719,6 +724,7 @@ def test_streaming_response_update(self, client: PromptFoundry) -> None:
719
724
],
720
725
name = "string" ,
721
726
parameters = {
727
+ "model_provider" : "OPENAI" ,
722
728
"model_name" : "string" ,
723
729
"response_format" : "JSON" ,
724
730
"temperature" : 0 ,
@@ -843,6 +849,7 @@ def test_path_params_update(self, client: PromptFoundry) -> None:
843
849
],
844
850
name = "string" ,
845
851
parameters = {
852
+ "model_provider" : "OPENAI" ,
846
853
"model_name" : "string" ,
847
854
"response_format" : "JSON" ,
848
855
"temperature" : 0 ,
@@ -1300,6 +1307,7 @@ async def test_method_create(self, async_client: AsyncPromptFoundry) -> None:
1300
1307
],
1301
1308
name = "string" ,
1302
1309
parameters = {
1310
+ "model_provider" : "OPENAI" ,
1303
1311
"model_name" : "string" ,
1304
1312
"response_format" : "JSON" ,
1305
1313
"temperature" : 0 ,
@@ -1416,6 +1424,7 @@ async def test_raw_response_create(self, async_client: AsyncPromptFoundry) -> No
1416
1424
],
1417
1425
name = "string" ,
1418
1426
parameters = {
1427
+ "model_provider" : "OPENAI" ,
1419
1428
"model_name" : "string" ,
1420
1429
"response_format" : "JSON" ,
1421
1430
"temperature" : 0 ,
@@ -1536,6 +1545,7 @@ async def test_streaming_response_create(self, async_client: AsyncPromptFoundry)
1536
1545
],
1537
1546
name = "string" ,
1538
1547
parameters = {
1548
+ "model_provider" : "OPENAI" ,
1539
1549
"model_name" : "string" ,
1540
1550
"response_format" : "JSON" ,
1541
1551
"temperature" : 0 ,
@@ -1659,6 +1669,7 @@ async def test_method_update(self, async_client: AsyncPromptFoundry) -> None:
1659
1669
],
1660
1670
name = "string" ,
1661
1671
parameters = {
1672
+ "model_provider" : "OPENAI" ,
1662
1673
"model_name" : "string" ,
1663
1674
"response_format" : "JSON" ,
1664
1675
"temperature" : 0 ,
@@ -1776,6 +1787,7 @@ async def test_raw_response_update(self, async_client: AsyncPromptFoundry) -> No
1776
1787
],
1777
1788
name = "string" ,
1778
1789
parameters = {
1790
+ "model_provider" : "OPENAI" ,
1779
1791
"model_name" : "string" ,
1780
1792
"response_format" : "JSON" ,
1781
1793
"temperature" : 0 ,
@@ -1897,6 +1909,7 @@ async def test_streaming_response_update(self, async_client: AsyncPromptFoundry)
1897
1909
],
1898
1910
name = "string" ,
1899
1911
parameters = {
1912
+ "model_provider" : "OPENAI" ,
1900
1913
"model_name" : "string" ,
1901
1914
"response_format" : "JSON" ,
1902
1915
"temperature" : 0 ,
@@ -2021,6 +2034,7 @@ async def test_path_params_update(self, async_client: AsyncPromptFoundry) -> Non
2021
2034
],
2022
2035
name = "string" ,
2023
2036
parameters = {
2037
+ "model_provider" : "OPENAI" ,
2024
2038
"model_name" : "string" ,
2025
2039
"response_format" : "JSON" ,
2026
2040
"temperature" : 0 ,
0 commit comments