Skip to content

Conversation

zhutao100
Copy link
Contributor

This PR enhances the token limits functionality by:

  1. Adding support for new models:

    • GLM-4.6 with 200k token limit
    • DeepSeek-v3.2-exp with 128k token limit
  2. Improving the model name normalization logic to:

    • Handle more date-like suffixes (e.g., deepseek-r1-0528)
    • Handle -exp suffixes in model names
    • Better distinguish between version numbers that should be stripped vs. kept
    • Improve handling of version numbers with and without the 'v' prefix

The changes are focused on improving the accuracy of token limit detection for various AI models, which is critical for ensuring proper handling of different model capabilities in the Qwen Code CLI.

Testing Matrix

🍏 🪟 🐧
npm run 👍
npx 👍
Docker
Podman - -
Seatbelt - -

Copy link
Collaborator

@Mingholy Mingholy left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@Mingholy Mingholy merged commit 9d66462 into QwenLM:main Oct 14, 2025
33 of 34 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants