Released Apr 7, 2026202,752 context$0.95/M input tokens$3.15/M output tokens
GLM-5.1 delivers a major leap in coding capability, with particularly significant gains in handling long-horizon tasks. Unlike previous models built around minute-level interactions, GLM-5.1 can work independently and continuously on a single task for more than 8 hours, autonomously planning, executing, and improving itself throughout the process, ultimately delivering complete, engineering-grade results.
Recent activity on GLM 5.1
Total usage per day on OpenRouter
Prompt
24.6B
Reasoning
276M
Completion
185M
Prompt tokens measure input size. Reasoning tokens show internal thinking before a response. Completion tokens reflect total output length.