feat(zai): update GLM-5 model specs with correct pricing and token limits#11479
Draft
roomote[bot] wants to merge 1 commit intomainfrom
Draft
feat(zai): update GLM-5 model specs with correct pricing and token limits#11479roomote[bot] wants to merge 1 commit intomainfrom
roomote[bot] wants to merge 1 commit intomainfrom
Conversation
2 tasks
Contributor
Author
Review complete -- no issues found. International GLM-5 pricing and token limits match the official docs. Mainland China pricing is proportionally scaled from GLM-4.7 ratios, which is consistent with the existing pattern. All 33 tests pass. Mention @roomote in a comment to request specific changes to this pull request or fix all unresolved issues. |
|
something is wrong with this bot roomote. 2 days for such an easy adjustment and still not promoted. |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Related GitHub Issue
Closes: #11438
Description
This PR attempts to address Issue #11438 by updating the GLM-5 model entries in the Z.ai provider with correct specs from the official docs, as provided by contributors in the issue comments.
The previous PR #11443 had merge conflicts, so this is a fresh implementation on current main.
International (z.ai) GLM-5 changes:
maxTokens: 16,384 -> 128,000 (128k max output per docs)contextWindow: 202,752 -> 200,000 (200k per docs)inputPrice: $0.60 -> $1.00 per 1M tokensoutputPrice: $2.20 -> $3.20 per 1M tokenscacheReadsPrice: $0.11 -> $0.20 per 1M tokensMainland China (bigmodel.cn) GLM-5 changes:
maxTokens: 16,384 -> 128,000contextWindow: 202,752 -> 200,000Features (already correctly configured, unchanged):
Feedback and guidance are welcome.
Test Procedure
npx vitest run api/providers/__tests__/zai.spec.tsfromsrc/directory -- all 33 tests passPre-Submission Checklist
Documentation Updates
Additional Notes
The GLM-5 model entry already existed in the codebase but with placeholder values copied from GLM-4.7. This PR corrects those values using the specs provided by @damianar1984 in the issue comments, sourced from the official z.ai documentation. This replaces the conflicting PR #11443 with a clean implementation on current main.