GLM-5 is Z.ai's MIT-licensed 744B-parameter MoE frontier model with 40B active per token, released February 11, 2026 — the first frontier-tier model trained end-to-end on Huawei Ascend silicon. It scores 77.8% on SWE-bench Verified at API pricing roughly 5–8x cheaper than Claude Opus 4.6.
GLM-5 is the flagship February 2026 frontier-class large language model from [[China/Z.ai|Z.ai]] (formerly Zhipu AI), and the first frontier-tier model trained end-to-end on Chinese-domestic Huawei Ascend silicon. Released February 11, 2026 under the MIT license, GLM-5 is a 744-billion-parameter Mixture-of-Experts model with 40 billion active parameters per token. Z.ai claims GLM-5 matches Claude Opus 4.5 and GPT-5.2 on coding and agent tasks, scoring 77.8% on SWE-bench Verified, 92.7% on AIME 2026, and 86.0% on GPQA-Diamond, while leading open-source models on BrowseComp, Vending Bench 2, and MCP-Atlas.
The release matters as much for what it proves about Chinese hardware as for its benchmark scores: GLM-5 was trained entirely on Huawei Ascend chips with the MindSpore framework, and inference runs across a fully domestic stack including Moore Threads, Cambricon, and Kunlunxin chips. Combined with API pricing of roughly $1.00 per million input tokens and $3.20 per million output tokens — approximately 5x cheaper than Claude Opus 4.6 on input and ~8x cheaper on output — GLM-5 sets a new floor for capable open-weight models worldwide.
Coding & SWE-bench Verified: 77.8% on SWE-bench Verified, in the same band as Claude Opus 4.5 and GPT-5.2 — the headline coding benchmark for production agentic software engineering.
Math Reasoning: 92.7% on AIME 2026, putting GLM-5 at the leading edge of contest-math performance among open-weight models.
Scientific Reasoning: 86.0% on GPQA-Diamond — the most demanding multiple-choice graduate-level science benchmark.
Agentic Workflows: Leads open-source models on BrowseComp (browser-driven task completion), Vending Bench 2 (multi-step business operations), and MCP-Atlas (Model Context Protocol agent benchmarks).
Multilinguality: Strong Chinese-language performance reflecting Z.ai's primary market, with competitive English-language results on standard evals.
Open-Weight Availability: Full weights released under MIT license, allowing self-hosted deployment without commercial-license restrictions.
GLM-5's benchmark claims are self-reported by Z.ai; while early independent replications support the core claims, comprehensive third-party evaluation is still ongoing as of May 2026. The 744B-parameter total footprint requires substantial GPU memory for self-hosted inference even with the 40B active-parameter MoE design — most enterprise users will continue to rely on the API. Geopolitically, U.S. enterprise customers may face procurement-policy or export-control complications when adopting GLM-5 given Z.ai's Chinese national-champion positioning and use of Huawei training hardware. Cybersecurity and content-safety alignment is calibrated to Chinese regulatory standards, which differ in some areas from Western norms.
May 8, 2026