zai-org/GLM-5
TEXT GENERATIONConcurrency Cost:4Model Size:754BQuant:FP8Ctx Length:32kPublished:Feb 11, 2026License:mitArchitecture:Transformer2.0K Open Weights Warm

GLM-5, developed by zai-org, is a large language model with 744 billion parameters (40 billion active) trained on 28.5 trillion tokens, designed for complex systems engineering and long-horizon agentic tasks. It integrates DeepSeek Sparse Attention (DSA) for reduced deployment cost and improved long-context capacity. The model excels in reasoning, coding, and agentic benchmarks, achieving best-in-class performance among open-source models by leveraging a novel asynchronous RL infrastructure called slime for efficient post-training.

Loading preview...