Skip to content

Commit bd6fecc

Browse files
feat(api): manual updates
1 parent 1f9bff4 commit bd6fecc

File tree

5 files changed

+29
-29
lines changed

5 files changed

+29
-29
lines changed

.stats.yml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
configured_endpoints: 76
2-
openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/digitalocean%2Fgradientai-e8b3cbc80e18e4f7f277010349f25e1319156704f359911dc464cc21a0d077a6.yml
1+
configured_endpoints: 77
2+
openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/digitalocean%2Fgradientai-391afaae764eb758523b67805cb47ae3bc319dc119d83414afdd66f123ceaf5c.yml
33
openapi_spec_hash: c773d792724f5647ae25a5ae4ccec208
4-
config_hash: 558ec54e9e056494abf623ff424c104e
4+
config_hash: 6edaff3557194ba8897d14f7ca74589c

README.md

Lines changed: 11 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -73,7 +73,7 @@ client = AsyncGradientAI(
7373

7474

7575
async def main() -> None:
76-
completion = await client.agents.chat.completions.create(
76+
completion = await client.chat.completions.create(
7777
messages=[
7878
{
7979
"role": "user",
@@ -115,7 +115,7 @@ async def main() -> None:
115115
api_key=os.environ.get("GRADIENTAI_API_KEY"), # This is the default and can be omitted
116116
http_client=DefaultAioHttpClient(),
117117
) as client:
118-
completion = await client.agents.chat.completions.create(
118+
completion = await client.chat.completions.create(
119119
messages=[
120120
{
121121
"role": "user",
@@ -139,7 +139,7 @@ from gradientai import GradientAI
139139

140140
client = GradientAI()
141141

142-
stream = client.agents.chat.completions.create(
142+
stream = client.chat.completions.create(
143143
messages=[
144144
{
145145
"role": "user",
@@ -160,7 +160,7 @@ from gradientai import AsyncGradientAI
160160

161161
client = AsyncGradientAI()
162162

163-
stream = await client.agents.chat.completions.create(
163+
stream = await client.chat.completions.create(
164164
messages=[
165165
{
166166
"role": "user",
@@ -192,7 +192,7 @@ from gradientai import GradientAI
192192

193193
client = GradientAI()
194194

195-
completion = client.agents.chat.completions.create(
195+
completion = client.chat.completions.create(
196196
messages=[
197197
{
198198
"content": "string",
@@ -221,7 +221,7 @@ from gradientai import GradientAI
221221
client = GradientAI()
222222

223223
try:
224-
client.agents.chat.completions.create(
224+
client.chat.completions.create(
225225
messages=[
226226
{
227227
"role": "user",
@@ -272,7 +272,7 @@ client = GradientAI(
272272
)
273273

274274
# Or, configure per-request:
275-
client.with_options(max_retries=5).agents.chat.completions.create(
275+
client.with_options(max_retries=5).chat.completions.create(
276276
messages=[
277277
{
278278
"role": "user",
@@ -303,7 +303,7 @@ client = GradientAI(
303303
)
304304

305305
# Override per-request:
306-
client.with_options(timeout=5.0).agents.chat.completions.create(
306+
client.with_options(timeout=5.0).chat.completions.create(
307307
messages=[
308308
{
309309
"role": "user",
@@ -352,7 +352,7 @@ The "raw" Response object can be accessed by prefixing `.with_raw_response.` to
352352
from gradientai import GradientAI
353353

354354
client = GradientAI()
355-
response = client.agents.chat.completions.with_raw_response.create(
355+
response = client.chat.completions.with_raw_response.create(
356356
messages=[{
357357
"role": "user",
358358
"content": "What is the capital of France?",
@@ -361,7 +361,7 @@ response = client.agents.chat.completions.with_raw_response.create(
361361
)
362362
print(response.headers.get('X-My-Header'))
363363

364-
completion = response.parse() # get the object that `agents.chat.completions.create()` would have returned
364+
completion = response.parse() # get the object that `chat.completions.create()` would have returned
365365
print(completion.choices)
366366
```
367367

@@ -376,7 +376,7 @@ The above interface eagerly reads the full response body when you make the reque
376376
To stream the response body, use `.with_streaming_response` instead, which requires a context manager and only reads the response body once you call `.read()`, `.text()`, `.json()`, `.iter_bytes()`, `.iter_text()`, `.iter_lines()` or `.parse()`. In the async client, these are async methods.
377377

378378
```python
379-
with client.agents.chat.completions.with_streaming_response.create(
379+
with client.chat.completions.with_streaming_response.create(
380380
messages=[
381381
{
382382
"role": "user",

api.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -70,7 +70,7 @@ from gradientai.types.agents.chat import CompletionCreateResponse
7070

7171
Methods:
7272

73-
- <code title="post /chat/completions">client.agents.chat.completions.<a href="./src/gradientai/resources/agents/chat/completions.py">create</a>(\*\*<a href="src/gradientai/types/agents/chat/completion_create_params.py">params</a>) -> <a href="./src/gradientai/types/agents/chat/completion_create_response.py">CompletionCreateResponse</a></code>
73+
- <code title="post /chat/completions?agent=true">client.agents.chat.completions.<a href="./src/gradientai/resources/agents/chat/completions.py">create</a>(\*\*<a href="src/gradientai/types/agents/chat/completion_create_params.py">params</a>) -> <a href="./src/gradientai/types/agents/chat/completion_create_response.py">CompletionCreateResponse</a></code>
7474

7575
## EvaluationMetrics
7676

src/gradientai/resources/agents/chat/completions.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -414,9 +414,9 @@ def create(
414414
timeout: float | httpx.Timeout | None | NotGiven = NOT_GIVEN,
415415
) -> CompletionCreateResponse | Stream[ChatCompletionChunk]:
416416
return self._post(
417-
"/chat/completions"
417+
"/chat/completions?agent=true"
418418
if self._client._base_url_overridden
419-
else "https://inference.do-ai.run/v1/chat/completions",
419+
else "https://inference.do-ai.run/v1/chat/completions?agent=true",
420420
body=maybe_transform(
421421
{
422422
"messages": messages,
@@ -838,9 +838,9 @@ async def create(
838838
timeout: float | httpx.Timeout | None | NotGiven = NOT_GIVEN,
839839
) -> CompletionCreateResponse | AsyncStream[ChatCompletionChunk]:
840840
return await self._post(
841-
"/chat/completions"
841+
"/chat/completions?agent=true"
842842
if self._client._base_url_overridden
843-
else "https://inference.do-ai.run/v1/chat/completions",
843+
else "https://inference.do-ai.run/v1/chat/completions?agent=true",
844844
body=await async_maybe_transform(
845845
{
846846
"messages": messages,

tests/test_client.py

Lines changed: 10 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -879,7 +879,7 @@ def test_retrying_timeout_errors_doesnt_leak(self, respx_mock: MockRouter, clien
879879
respx_mock.post("/chat/completions").mock(side_effect=httpx.TimeoutException("Test timeout error"))
880880

881881
with pytest.raises(APITimeoutError):
882-
client.agents.chat.completions.with_streaming_response.create(
882+
client.chat.completions.with_streaming_response.create(
883883
messages=[
884884
{
885885
"content": "string",
@@ -897,7 +897,7 @@ def test_retrying_status_errors_doesnt_leak(self, respx_mock: MockRouter, client
897897
respx_mock.post("/chat/completions").mock(return_value=httpx.Response(500))
898898

899899
with pytest.raises(APIStatusError):
900-
client.agents.chat.completions.with_streaming_response.create(
900+
client.chat.completions.with_streaming_response.create(
901901
messages=[
902902
{
903903
"content": "string",
@@ -934,7 +934,7 @@ def retry_handler(_request: httpx.Request) -> httpx.Response:
934934

935935
respx_mock.post("/chat/completions").mock(side_effect=retry_handler)
936936

937-
response = client.agents.chat.completions.with_raw_response.create(
937+
response = client.chat.completions.with_raw_response.create(
938938
messages=[
939939
{
940940
"content": "string",
@@ -966,7 +966,7 @@ def retry_handler(_request: httpx.Request) -> httpx.Response:
966966

967967
respx_mock.post("/chat/completions").mock(side_effect=retry_handler)
968968

969-
response = client.agents.chat.completions.with_raw_response.create(
969+
response = client.chat.completions.with_raw_response.create(
970970
messages=[
971971
{
972972
"content": "string",
@@ -998,7 +998,7 @@ def retry_handler(_request: httpx.Request) -> httpx.Response:
998998

999999
respx_mock.post("/chat/completions").mock(side_effect=retry_handler)
10001000

1001-
response = client.agents.chat.completions.with_raw_response.create(
1001+
response = client.chat.completions.with_raw_response.create(
10021002
messages=[
10031003
{
10041004
"content": "string",
@@ -1888,7 +1888,7 @@ async def test_retrying_timeout_errors_doesnt_leak(
18881888
respx_mock.post("/chat/completions").mock(side_effect=httpx.TimeoutException("Test timeout error"))
18891889

18901890
with pytest.raises(APITimeoutError):
1891-
await async_client.agents.chat.completions.with_streaming_response.create(
1891+
await async_client.chat.completions.with_streaming_response.create(
18921892
messages=[
18931893
{
18941894
"content": "string",
@@ -1908,7 +1908,7 @@ async def test_retrying_status_errors_doesnt_leak(
19081908
respx_mock.post("/chat/completions").mock(return_value=httpx.Response(500))
19091909

19101910
with pytest.raises(APIStatusError):
1911-
await async_client.agents.chat.completions.with_streaming_response.create(
1911+
await async_client.chat.completions.with_streaming_response.create(
19121912
messages=[
19131913
{
19141914
"content": "string",
@@ -1946,7 +1946,7 @@ def retry_handler(_request: httpx.Request) -> httpx.Response:
19461946

19471947
respx_mock.post("/chat/completions").mock(side_effect=retry_handler)
19481948

1949-
response = await client.agents.chat.completions.with_raw_response.create(
1949+
response = await client.chat.completions.with_raw_response.create(
19501950
messages=[
19511951
{
19521952
"content": "string",
@@ -1979,7 +1979,7 @@ def retry_handler(_request: httpx.Request) -> httpx.Response:
19791979

19801980
respx_mock.post("/chat/completions").mock(side_effect=retry_handler)
19811981

1982-
response = await client.agents.chat.completions.with_raw_response.create(
1982+
response = await client.chat.completions.with_raw_response.create(
19831983
messages=[
19841984
{
19851985
"content": "string",
@@ -2012,7 +2012,7 @@ def retry_handler(_request: httpx.Request) -> httpx.Response:
20122012

20132013
respx_mock.post("/chat/completions").mock(side_effect=retry_handler)
20142014

2015-
response = await client.agents.chat.completions.with_raw_response.create(
2015+
response = await client.chat.completions.with_raw_response.create(
20162016
messages=[
20172017
{
20182018
"content": "string",

0 commit comments

Comments
 (0)