Skip to content

Zhipu AI Unveils GLM-4.6: Bigger Context Window, Improved Efficiency

GLM-4.6's expanded context window is a game-changer for coding and long-form text generation. Its improved efficiency makes it a standout choice for tasks requiring extended input.

This picture contains a box which is in red, orange and blue color. On the top of the box, we see a...
This picture contains a box which is in red, orange and blue color. On the top of the box, we see a robot and text written as "AUTOBOT TRACKS". In the background, it is black in color and it is blurred.

Zhipu AI Unveils GLM-4.6: Bigger Context Window, Improved Efficiency

Zhipu AI has launched GLM-4.6, the latest iteration of its GLM series, boasting significant improvements and a larger context window. The model aims to reduce token consumption in practical tasks and is licensed under the MIT license.

GLM-4.6 offers a substantial increase in its input window, now supporting 200K tokens, double that of its predecessor GLM-4.5. This expansion allows the model to excel in tasks requiring extended input, such as coding and long-form text generation.

On the extended CC-Bench, GLM-4.6 performs near parity with Claude Sonnet 4, a model renowned for its high performance. Notably, GLM-4.6 uses 15% fewer tokens than GLM-4.5, indicating improved efficiency. The model has 355B parameters and supports BF16/F32 tensors.

GLM-4.6 is available via Z.ai API and OpenRouter, making it accessible for integration with popular coding agents. It also supports local serving with vLLM and SGLang, providing flexibility for users. Weights for the model are openly available on Hugging Face and ModelScope, enabling local deployment.

Zhipu AI's GLM-4.6 is a significant update to the GLM series, offering a larger context window and improved performance. With open weights and integration capabilities, it is poised to make a substantial impact in tasks requiring extended input and efficient token consumption.

Read also:

Latest