LATEST MODEL

GLM-5

Z.ai 🧠 The Thinker Released February 2026

Zhipu AI's open-weight 744B MoE foundation model purpose-built for agentic engineering, trained entirely on Huawei Ascend chips

GLM-5

Z.aiFebruary 2026

Latest

Training Data

28.5 trillion tokens up to late 2025

GLM-5

February 2026

Parameters

744 billion (40B active)

Training Method

MoE with MLA and DeepSeek Sparse Attention on Huawei Ascend 910B chips

Context Window

200,000 tokens

Knowledge Cutoff

Not disclosed

Key Features

MIT-Licensed 744B MoE • DeepSeek Sparse Attention • Trained on Huawei Ascend 910B

Capabilities

Reasoning: Excellent

Coding: Excellent

Agentic Tasks: Very Good

What's New in This Version

Doubles parameters from GLM-4.5's 355B to 744B with 28.5T training tokens and reduces hallucination rate from 90% to 34% through asynchronous RL

Zhipu AI's open-weight 744B MoE foundation model purpose-built for agentic engineering, trained entirely on Huawei Ascend chips

What's New in This Version

Doubles parameters from GLM-4.5's 355B to 744B with 28.5T training tokens and reduces hallucination rate from 90% to 34% through asynchronous RL

Technical Specifications

Parameters 744 billion (40B active)
Context Window 200,000 tokens
Training Method MoE with MLA and DeepSeek Sparse Attention on Huawei Ascend 910B chips
Knowledge Cutoff Not disclosed
Training Data 28.5 trillion tokens up to late 2025

Key Features

MIT-Licensed 744B MoE DeepSeek Sparse Attention Trained on Huawei Ascend 910B

Capabilities

Reasoning: Excellent
Coding: Excellent
Agentic Tasks: Very Good
Theme
Language
Support
© funclosure 2025