Skip to content

Zhipu AI Unveils GLM-4.6: A Giant Leap in Large Language Models

GLM-4.6 boasts a massive 200K token input window and 15% reduced token consumption. It's set to revolutionize large language model applications.

This picture contains a box which is in red, orange and blue color. On the top of the box, we see a...
This picture contains a box which is in red, orange and blue color. On the top of the box, we see a robot and text written as "AUTOBOT TRACKS". In the background, it is black in color and it is blurred.

Zhipu AI Unveils GLM-4.6: A Giant Leap in Large Language Models

Zhipu AI has unveiled GLM-4.6, the latest iteration in its GLM series. This new version focuses on agentic workflows and long-context reasoning, boasting 355B parameters and supporting local serving with vLLM and SGLang.

GLM-4.6 shows significant improvements over its predecessor, GLM-4.5, demonstrating clear gains across eight public benchmarks. It also performs on par with Claude Sonnet 4/4.6 on several tasks. The model's input window stretches to 200K tokens, with a maximum output of 128K tokens.

In terms of licensing and availability, GLM-4.6 is released under the MIT license. Its weights are accessible on both Hugging Face and ModelScope, enabling open deployment. Additionally, the model can be integrated with popular coding agents and is available via Z.ai API and OpenRouter. Zhipu AI has also targeted lower token consumption in applied tasks, with GLM-4.6 using 15% fewer tokens than GLM-4.5 on the extended CC-Bench, while approaching parity with Claude Sonnet 4.

GLM-4.6, with its enhanced capabilities and open deployment options, is set to advance the field of large language models. Its improved performance and efficiency make it a compelling choice for a variety of applications.

Read also:

Latest