Skip to main content

Models

2025-01-14
GLM-Image
  • We’ve launched GLM-Image, a state-of-the-art image generation model built on a multimodal architecture and fully trained on domestic chips, combining autoregressive semantic understanding with diffusion-based decoding to deliver high-quality, controllable visual generation.
  • The release significantly enhances performance in knowledge-intensive scenarios, with more stable and accurate text rendering inside images, making GLM-Image especially well suited for commercial design, educational illustrations, and content-rich visual applications.Learn more in our documentation.*
2025-12-22
GLM-4.7
  • We’ve released GLM-4.7, our latest flagship foundation model with significant improvements in coding, reasoning, and agentic capabilities. It delivers more reliable code generation, stronger long-context understanding, and improved end-to-end task execution across real-world development workflows.
  • The update brings open-source SOTA performance on major coding and reasoning benchmarks, enhanced agentic coding for goal-driven, multi-step tasks, and improved front-end and document generation quality. Learn more in our documentation.*
2025-12-11
AutoGLM-Phone-Multilingual
  • We’ve launched AutoGLM-Phone-Multilingual, our latest multimodal mobile automation framework that understands screen content and executes real actions through ADB. It enables natural-language task execution across 50+ mainstream apps, delivering true end-to-end mobile control.
  • The update introduces multilingual support (English & Chinese), enhanced workflow planning capabilities, and improved task execution reliability. Learn more in our documentation.*
2025-12-10
GLM-ASR-2512
  • We’ve launched GLM-ASR-2512, our ASR model, delivering industry-leading accuracy with a Character Error Rate of just 0.0717, and significantly improved performance across real-world multilingual and accent-rich scenarios.
  • The update introduces enhanced custom dictionary support and expanded specialized terminology recognition. Learn more in our documentation.*
2025-12-08
GLM-4.6V
  • We’re excited to introduce GLM-4.6V, Z.ai’s latest iteration in multimodal large language models. This version enhances vision understanding, achieving state-of-the-art performance in tasks involving images and text.
  • The update also expands the context window to 128K, enabling more efficient processing of long inputs and complex multimodal tasks. Learn more in our documentation.*
2025-09-30
GLM-4.6
  • We’ve launched GLM-4.6, the flagship coding model, showcasing enhanced performance in both public benchmarks and real-world programming tasks, making it the leading coding model in China.
  • The update also expands the context window to 200K, improving its ability to handle longer code and complex agent tasks. Learn more in our documentation.*
2025-08-11
GLM-4.5V
  • We’ve launched GLM-4.5V, a 100B-scale open-source vision reasoning model, supporting a broad range of visual tasks including video understanding, visual grounding, GUI agents and etc.
  • The update also adds a new thinking mode. Learn more in our documentation.*
2025-08-08
GLM Slide/Poster Agent(beta)
  • We’ve launched GLM Slide/Poster Agent, an AI-powered creation agent that combines information retrieval, content structuring, and visual layout design to generate professional-grade slides and posters from natural language instructions.
  • The update also brings a seamless integration of content generation with design conventions. Learn more in our documentation.*
2025-07-28
GLM-4.5 Series
  • We’ve launched GLM-4.5, our latest native agentic LLM, delivering doubled parameter efficiency and strong reasoning, coding, and agentic capabilities.
  • It also offers seamless one-click compatibility with the Claude Code framework. Learn more in our documentation.*
2025-07-15
CogVideoX-3
  • We’ve launched CogVideoX-3, an incremental upgrade to our video generation model with improved quality and new features.
  • It adds support for start and end frame synthesis. Learn more in our documentation.*