kennedyantonio0301/Affine-Tensor-h2-5D4Ug3BeJtaHm2D1vypjfCKnQQXt3VXzajyGjk2gSW269axP
TEXT GENERATIONConcurrency Cost:1Model Size:14BQuant:FP8Ctx Length:32kPublished:Jan 31, 2026License:mitArchitecture:Transformer Open Weights Cold

GLM-4.7-Flash is a 30 billion parameter Mixture-of-Experts (MoE) model developed by the GLM Team, designed to be the strongest in its class for lightweight deployment. It balances performance and efficiency, excelling particularly in agentic tasks, reasoning, and coding benchmarks. This model offers a new option for applications requiring high performance within a constrained resource environment.

Loading preview...