Models
- We’ve launched GLM-Image, a state-of-the-art image generation model built on a multimodal architecture and fully trained on domestic chips, combining autoregressive semantic understanding with diffusion-based decoding to deliver high-quality, controllable visual generation.
- The release significantly enhances performance in knowledge-intensive scenarios, with more stable and accurate text rendering inside images, making GLM-Image especially well suited for commercial design, educational illustrations, and content-rich visual applications.Learn more in our documentation.*
- We’ve released GLM-4.7, our latest flagship foundation model with significant improvements in coding, reasoning, and agentic capabilities. It delivers more reliable code generation, stronger long-context understanding, and improved end-to-end task execution across real-world development workflows.
- The update brings open-source SOTA performance on major coding and reasoning benchmarks, enhanced agentic coding for goal-driven, multi-step tasks, and improved front-end and document generation quality. Learn more in our documentation.*
- We’ve launched AutoGLM-Phone-Multilingual, our latest multimodal mobile automation framework that understands screen content and executes real actions through ADB. It enables natural-language task execution across 50+ mainstream apps, delivering true end-to-end mobile control.
- The update introduces multilingual support (English & Chinese), enhanced workflow planning capabilities, and improved task execution reliability. Learn more in our documentation.*
- We’ve launched GLM-ASR-2512, our ASR model, delivering industry-leading accuracy with a Character Error Rate of just 0.0717, and significantly improved performance across real-world multilingual and accent-rich scenarios.
- The update introduces enhanced custom dictionary support and expanded specialized terminology recognition. Learn more in our documentation.*
- We’re excited to introduce GLM-4.6V, Z.ai’s latest iteration in multimodal large language models. This version enhances vision understanding, achieving state-of-the-art performance in tasks involving images and text.
- The update also expands the context window to 128K, enabling more efficient processing of long inputs and complex multimodal tasks. Learn more in our documentation.*
- We’ve launched GLM-4.6, the flagship coding model, showcasing enhanced performance in both public benchmarks and real-world programming tasks, making it the leading coding model in China.
- The update also expands the context window to 200K, improving its ability to handle longer code and complex agent tasks. Learn more in our documentation.*
- We’ve launched GLM-4.5V, a 100B-scale open-source vision reasoning model, supporting a broad range of visual tasks including video understanding, visual grounding, GUI agents and etc.
- The update also adds a new thinking mode. Learn more in our documentation.*
- We’ve launched GLM Slide/Poster Agent, an AI-powered creation agent that combines information retrieval, content structuring, and visual layout design to generate professional-grade slides and posters from natural language instructions.
- The update also brings a seamless integration of content generation with design conventions. Learn more in our documentation.*
- We’ve launched GLM-4.5, our latest native agentic LLM, delivering doubled parameter efficiency and strong reasoning, coding, and agentic capabilities.
- It also offers seamless one-click compatibility with the Claude Code framework. Learn more in our documentation.*
- We’ve launched CogVideoX-3, an incremental upgrade to our video generation model with improved quality and new features.
- It adds support for start and end frame synthesis. Learn more in our documentation.*