Skip to content

Commit 7b3aa00

Browse files
Merge pull request #40 from gleanwork/speakeasy-sdk-regen-1749454024
chore: 🐝 Update SDK - Generate 0.6.4
2 parents 3f8fd5d + a3049b1 commit 7b3aa00

File tree

13 files changed

+75
-33
lines changed

13 files changed

+75
-33
lines changed

.speakeasy/gen.lock

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,12 @@
11
lockVersion: 2.0.0
22
id: 3e3290ca-0ee8-4981-b1bc-14536048fa63
33
management:
4-
docChecksum: 895c9d213122353173d7c129b2c8d4b7
4+
docChecksum: 1477738232aeb60aace8340880be72ae
55
docVersion: 0.9.0
6-
speakeasyVersion: 1.557.0
7-
generationVersion: 2.623.0
8-
releaseVersion: 0.6.3
9-
configChecksum: f4420151b362f554dd36cc2e52ba1688
6+
speakeasyVersion: 1.557.1
7+
generationVersion: 2.623.2
8+
releaseVersion: 0.6.4
9+
configChecksum: 97a162cbd1df1380dc8322bed97ec589
1010
repoURL: https://github.com/gleanwork/api-client-python.git
1111
installationURL: https://github.com/gleanwork/api-client-python.git
1212
published: true

.speakeasy/gen.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@ generation:
2525
generateNewTests: true
2626
skipResponseBodyAssertions: true
2727
python:
28-
version: 0.6.3
28+
version: 0.6.4
2929
additionalDependencies:
3030
dev: {}
3131
main: {}

.speakeasy/glean-merged-spec.yaml

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -7259,6 +7259,9 @@ components:
72597259
applicationId:
72607260
type: string
72617261
description: The ID of the application this request originates from, used to determine the configuration of underlying chat processes. This should correspond to the ID set during admin setup. If not specified, the default chat experience will be used.
7262+
agentId:
7263+
type: string
7264+
description: The ID of the Agent that should process this chat request. Only Agents with trigger set to 'User chat message' are invokable through this API. If not specified, the default chat experience will be used.
72627265
stream:
72637266
type: boolean
72647267
description: If set, response lines will be streamed one-by-one as they become available. Each will be a ChatResponse, formatted as JSON, and separated by a new line. If false, the entire response will be returned at once. Note that if this is set and the model being used does not support streaming, the model's response will not be streamed, but other messages from the endpoint still will be.

.speakeasy/workflow.lock

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,12 @@
1-
speakeasyVersion: 1.557.0
1+
speakeasyVersion: 1.557.1
22
sources:
33
Glean API:
44
sourceNamespace: glean-api-specs
5-
sourceRevisionDigest: sha256:d7ce1ceb600711b8e88b818186b93b6e4275f01811996d436752410c100339a4
6-
sourceBlobDigest: sha256:e56c126a42ae3d72395b2891ac56b8f5264993a298d9d4f5e68748395cdd54cf
5+
sourceRevisionDigest: sha256:254ed74184de2e725a0dc2cf521b132b29ece9d3ba6b12fd7e7b1ceaa2e6d043
6+
sourceBlobDigest: sha256:f2a59d8beefedbda0e59b25d09a930e1b907145f784bf8961fc31b862163d964
77
tags:
88
- latest
9-
- speakeasy-sdk-regen-1749247136
9+
- speakeasy-sdk-regen-1749454024
1010
Glean Client API:
1111
sourceNamespace: glean-client-api
1212
sourceRevisionDigest: sha256:4edc63ad559e4f2c9fb9ebf5edaaaaa9269f1874d271cfd84b441d6dacac43d2
@@ -17,10 +17,10 @@ targets:
1717
glean:
1818
source: Glean API
1919
sourceNamespace: glean-api-specs
20-
sourceRevisionDigest: sha256:d7ce1ceb600711b8e88b818186b93b6e4275f01811996d436752410c100339a4
21-
sourceBlobDigest: sha256:e56c126a42ae3d72395b2891ac56b8f5264993a298d9d4f5e68748395cdd54cf
20+
sourceRevisionDigest: sha256:254ed74184de2e725a0dc2cf521b132b29ece9d3ba6b12fd7e7b1ceaa2e6d043
21+
sourceBlobDigest: sha256:f2a59d8beefedbda0e59b25d09a930e1b907145f784bf8961fc31b862163d964
2222
codeSamplesNamespace: glean-api-specs-python-code-samples
23-
codeSamplesRevisionDigest: sha256:616f210191e4d9c7270884cda5aee3e53caa4a3c1eba93997f725bdd345eed8c
23+
codeSamplesRevisionDigest: sha256:bd2c8dd3d421bb8899be6f133e8dcd287ceaeb9e7022ad1b5f3415d7b335c1ab
2424
workflow:
2525
workflowVersion: 1.0.0
2626
speakeasyVersion: latest

RELEASES.md

Lines changed: 11 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -138,4 +138,14 @@ Based on:
138138
### Generated
139139
- [python v0.6.3] .
140140
### Releases
141-
- [PyPI v0.6.3] https://pypi.org/project/glean/0.6.3 - .
141+
- [PyPI v0.6.3] https://pypi.org/project/glean/0.6.3 - .
142+
143+
## 2025-06-09 07:26:40
144+
### Changes
145+
Based on:
146+
- OpenAPI Doc
147+
- Speakeasy CLI 1.557.1 (2.623.2) https://github.com/speakeasy-api/speakeasy
148+
### Generated
149+
- [python v0.6.4] .
150+
### Releases
151+
- [PyPI v0.6.4] https://pypi.org/project/glean/0.6.4 - .

docs/models/chatrequest.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -14,4 +14,5 @@
1414
| `timeout_millis` | *Optional[int]* | :heavy_minus_sign: | Timeout in milliseconds for the request. A `408` error will be returned if handling the request takes longer. | 30000 |
1515
| `session_info` | [Optional[models.SessionInfo]](../models/sessioninfo.md) | :heavy_minus_sign: | N/A | |
1616
| `application_id` | *Optional[str]* | :heavy_minus_sign: | The ID of the application this request originates from, used to determine the configuration of underlying chat processes. This should correspond to the ID set during admin setup. If not specified, the default chat experience will be used. | |
17+
| `agent_id` | *Optional[str]* | :heavy_minus_sign: | The ID of the Agent that should process this chat request. Only Agents with trigger set to 'User chat message' are invokable through this API. If not specified, the default chat experience will be used. | |
1718
| `stream` | *Optional[bool]* | :heavy_minus_sign: | If set, response lines will be streamed one-by-one as they become available. Each will be a ChatResponse, formatted as JSON, and separated by a new line. If false, the entire response will be returned at once. Note that if this is set and the model being used does not support streaming, the model's response will not be streamed, but other messages from the endpoint still will be. | |

0 commit comments

Comments
 (0)