Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion .stats.yml
Original file line number Diff line number Diff line change
@@ -1,2 +1,2 @@
configured_endpoints: 29
openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/writerai%2Fwriter-5d57be55ac24a66e8c78594a3a964a90f12f98fc51f8c758891ec7fee2b74c9f.yml
openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/writerai%2Fwriter-710e8a17bc916b755685592b3831e6732f3ba02904f970084aef0ac86fd79ed5.yml
40 changes: 20 additions & 20 deletions tests/api_resources/test_chat.py
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ class TestChat:
def test_method_chat_overload_1(self, client: Writer) -> None:
chat = client.chat.chat(
messages=[{"role": "user"}],
model="palmyra-x-004",
model="model",
)
assert_matches_type(ChatCompletion, chat, path=["response"])

Expand All @@ -31,7 +31,7 @@ def test_method_chat_with_all_params_overload_1(self, client: Writer) -> None:
messages=[
{
"role": "user",
"content": "Write a memo summarizing this earnings report.",
"content": "content",
"graph_data": {
"sources": [
{
Expand Down Expand Up @@ -69,7 +69,7 @@ def test_method_chat_with_all_params_overload_1(self, client: Writer) -> None:
],
}
],
model="palmyra-x-004",
model="model",
logprobs=True,
max_tokens=0,
n=0,
Expand All @@ -96,7 +96,7 @@ def test_method_chat_with_all_params_overload_1(self, client: Writer) -> None:
def test_raw_response_chat_overload_1(self, client: Writer) -> None:
response = client.chat.with_raw_response.chat(
messages=[{"role": "user"}],
model="palmyra-x-004",
model="model",
)

assert response.is_closed is True
Expand All @@ -108,7 +108,7 @@ def test_raw_response_chat_overload_1(self, client: Writer) -> None:
def test_streaming_response_chat_overload_1(self, client: Writer) -> None:
with client.chat.with_streaming_response.chat(
messages=[{"role": "user"}],
model="palmyra-x-004",
model="model",
) as response:
assert not response.is_closed
assert response.http_request.headers.get("X-Stainless-Lang") == "python"
Expand All @@ -122,7 +122,7 @@ def test_streaming_response_chat_overload_1(self, client: Writer) -> None:
def test_method_chat_overload_2(self, client: Writer) -> None:
chat_stream = client.chat.chat(
messages=[{"role": "user"}],
model="palmyra-x-004",
model="model",
stream=True,
)
chat_stream.response.close()
Expand All @@ -133,7 +133,7 @@ def test_method_chat_with_all_params_overload_2(self, client: Writer) -> None:
messages=[
{
"role": "user",
"content": "Write a memo summarizing this earnings report.",
"content": "content",
"graph_data": {
"sources": [
{
Expand Down Expand Up @@ -171,7 +171,7 @@ def test_method_chat_with_all_params_overload_2(self, client: Writer) -> None:
],
}
],
model="palmyra-x-004",
model="model",
stream=True,
logprobs=True,
max_tokens=0,
Expand All @@ -198,7 +198,7 @@ def test_method_chat_with_all_params_overload_2(self, client: Writer) -> None:
def test_raw_response_chat_overload_2(self, client: Writer) -> None:
response = client.chat.with_raw_response.chat(
messages=[{"role": "user"}],
model="palmyra-x-004",
model="model",
stream=True,
)

Expand All @@ -210,7 +210,7 @@ def test_raw_response_chat_overload_2(self, client: Writer) -> None:
def test_streaming_response_chat_overload_2(self, client: Writer) -> None:
with client.chat.with_streaming_response.chat(
messages=[{"role": "user"}],
model="palmyra-x-004",
model="model",
stream=True,
) as response:
assert not response.is_closed
Expand All @@ -229,7 +229,7 @@ class TestAsyncChat:
async def test_method_chat_overload_1(self, async_client: AsyncWriter) -> None:
chat = await async_client.chat.chat(
messages=[{"role": "user"}],
model="palmyra-x-004",
model="model",
)
assert_matches_type(ChatCompletion, chat, path=["response"])

Expand All @@ -239,7 +239,7 @@ async def test_method_chat_with_all_params_overload_1(self, async_client: AsyncW
messages=[
{
"role": "user",
"content": "Write a memo summarizing this earnings report.",
"content": "content",
"graph_data": {
"sources": [
{
Expand Down Expand Up @@ -277,7 +277,7 @@ async def test_method_chat_with_all_params_overload_1(self, async_client: AsyncW
],
}
],
model="palmyra-x-004",
model="model",
logprobs=True,
max_tokens=0,
n=0,
Expand All @@ -304,7 +304,7 @@ async def test_method_chat_with_all_params_overload_1(self, async_client: AsyncW
async def test_raw_response_chat_overload_1(self, async_client: AsyncWriter) -> None:
response = await async_client.chat.with_raw_response.chat(
messages=[{"role": "user"}],
model="palmyra-x-004",
model="model",
)

assert response.is_closed is True
Expand All @@ -316,7 +316,7 @@ async def test_raw_response_chat_overload_1(self, async_client: AsyncWriter) ->
async def test_streaming_response_chat_overload_1(self, async_client: AsyncWriter) -> None:
async with async_client.chat.with_streaming_response.chat(
messages=[{"role": "user"}],
model="palmyra-x-004",
model="model",
) as response:
assert not response.is_closed
assert response.http_request.headers.get("X-Stainless-Lang") == "python"
Expand All @@ -330,7 +330,7 @@ async def test_streaming_response_chat_overload_1(self, async_client: AsyncWrite
async def test_method_chat_overload_2(self, async_client: AsyncWriter) -> None:
chat_stream = await async_client.chat.chat(
messages=[{"role": "user"}],
model="palmyra-x-004",
model="model",
stream=True,
)
await chat_stream.response.aclose()
Expand All @@ -341,7 +341,7 @@ async def test_method_chat_with_all_params_overload_2(self, async_client: AsyncW
messages=[
{
"role": "user",
"content": "Write a memo summarizing this earnings report.",
"content": "content",
"graph_data": {
"sources": [
{
Expand Down Expand Up @@ -379,7 +379,7 @@ async def test_method_chat_with_all_params_overload_2(self, async_client: AsyncW
],
}
],
model="palmyra-x-004",
model="model",
stream=True,
logprobs=True,
max_tokens=0,
Expand All @@ -406,7 +406,7 @@ async def test_method_chat_with_all_params_overload_2(self, async_client: AsyncW
async def test_raw_response_chat_overload_2(self, async_client: AsyncWriter) -> None:
response = await async_client.chat.with_raw_response.chat(
messages=[{"role": "user"}],
model="palmyra-x-004",
model="model",
stream=True,
)

Expand All @@ -418,7 +418,7 @@ async def test_raw_response_chat_overload_2(self, async_client: AsyncWriter) ->
async def test_streaming_response_chat_overload_2(self, async_client: AsyncWriter) -> None:
async with async_client.chat.with_streaming_response.chat(
messages=[{"role": "user"}],
model="palmyra-x-004",
model="model",
stream=True,
) as response:
assert not response.is_closed
Expand Down
12 changes: 6 additions & 6 deletions tests/test_client.py
Original file line number Diff line number Diff line change
Expand Up @@ -787,7 +787,7 @@ def retry_handler(_request: httpx.Request) -> httpx.Response:

respx_mock.post("/v1/chat").mock(side_effect=retry_handler)

response = client.chat.with_raw_response.chat(messages=[{"role": "user"}], model="palmyra-x-004")
response = client.chat.with_raw_response.chat(messages=[{"role": "user"}], model="model")

assert response.retries_taken == failures_before_success
assert int(response.http_request.headers.get("x-stainless-retry-count")) == failures_before_success
Expand All @@ -812,7 +812,7 @@ def retry_handler(_request: httpx.Request) -> httpx.Response:
respx_mock.post("/v1/chat").mock(side_effect=retry_handler)

response = client.chat.with_raw_response.chat(
messages=[{"role": "user"}], model="palmyra-x-004", extra_headers={"x-stainless-retry-count": Omit()}
messages=[{"role": "user"}], model="model", extra_headers={"x-stainless-retry-count": Omit()}
)

assert len(response.http_request.headers.get_list("x-stainless-retry-count")) == 0
Expand All @@ -837,7 +837,7 @@ def retry_handler(_request: httpx.Request) -> httpx.Response:
respx_mock.post("/v1/chat").mock(side_effect=retry_handler)

response = client.chat.with_raw_response.chat(
messages=[{"role": "user"}], model="palmyra-x-004", extra_headers={"x-stainless-retry-count": "42"}
messages=[{"role": "user"}], model="model", extra_headers={"x-stainless-retry-count": "42"}
)

assert response.http_request.headers.get("x-stainless-retry-count") == "42"
Expand Down Expand Up @@ -1590,7 +1590,7 @@ def retry_handler(_request: httpx.Request) -> httpx.Response:

respx_mock.post("/v1/chat").mock(side_effect=retry_handler)

response = await client.chat.with_raw_response.chat(messages=[{"role": "user"}], model="palmyra-x-004")
response = await client.chat.with_raw_response.chat(messages=[{"role": "user"}], model="model")

assert response.retries_taken == failures_before_success
assert int(response.http_request.headers.get("x-stainless-retry-count")) == failures_before_success
Expand All @@ -1616,7 +1616,7 @@ def retry_handler(_request: httpx.Request) -> httpx.Response:
respx_mock.post("/v1/chat").mock(side_effect=retry_handler)

response = await client.chat.with_raw_response.chat(
messages=[{"role": "user"}], model="palmyra-x-004", extra_headers={"x-stainless-retry-count": Omit()}
messages=[{"role": "user"}], model="model", extra_headers={"x-stainless-retry-count": Omit()}
)

assert len(response.http_request.headers.get_list("x-stainless-retry-count")) == 0
Expand All @@ -1642,7 +1642,7 @@ def retry_handler(_request: httpx.Request) -> httpx.Response:
respx_mock.post("/v1/chat").mock(side_effect=retry_handler)

response = await client.chat.with_raw_response.chat(
messages=[{"role": "user"}], model="palmyra-x-004", extra_headers={"x-stainless-retry-count": "42"}
messages=[{"role": "user"}], model="model", extra_headers={"x-stainless-retry-count": "42"}
)

assert response.http_request.headers.get("x-stainless-retry-count") == "42"
Expand Down