feat(zai): update GLM-5 model specs with correct pricing and token limits #11443
+14
−14
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Related GitHub Issue
Closes: #11438
Description
This PR attempts to address Issue #11438 by updating the GLM-5 model entry in the Z.ai provider with correct specs as provided in the issue comments (from z.ai docs).
Changes made:
International (z.ai) GLM-5:
maxTokens: 16,384 -> 128,000 (128k max output per docs)contextWindow: 202,752 -> 200,000 (200k per docs)inputPrice: $0.60 -> $1.00 per 1M tokensoutputPrice: $2.20 -> $3.20 per 1M tokenscacheReadsPrice: $0.11 -> $0.20 per 1M tokensMainland China (bigmodel.cn) GLM-5:
maxTokens: 16,384 -> 128,000contextWindow: 202,752 -> 200,000Other features (already correctly configured):
Feedback and guidance are welcome.
Test Procedure
npx vitest run api/providers/__tests__/zai.spec.tsfromsrc/directory -- all 31 tests passPre-Submission Checklist
Documentation Updates
Additional Notes
The GLM-5 model entry already existed in the codebase but with placeholder values copied from GLM-4.7. This PR corrects those values using the specs provided by @damianar1984 in the issue comments, sourced from the official z.ai documentation.