All models

Z.AI: GLM 4.6

by Z.AI

Compared with GLM-4.5, this generation brings several key improvements: Longer context window: The context window has been expanded from 128K to 200K tokens, enabling the model to handle more complex agentic tasks. Superior coding performance: The model achieves higher scores on code benchmarks and demonstrates better real-world performance in applications such as Claude Code、Cline、Roo Code and Kilo Code, including improvements in generating visually polished front-end pages. Advanced reasoning: GLM-4.6 shows a clear improvement in reasoning performance and supports tool use during inference, leading to stronger overall capability. More capable agents: GLM-4.6 exhibits stronger performance in tool using and search-based agents, and integrates more effectively within agent frameworks. Refined writing: Better aligns with human preferences in style and readability, and performs more naturally in role-playing scenarios.

Avg Score

95.0%

21 answers

Avg Latency

38.3s

9 runs

Pricing

$0.44

input

/

$1.76

output

per 1M tokens

Context

205K

tokens

Alternatives

Models with similar or better quality but different tradeoffs

Same Quality, Cheaper

Models with similar or better performance at a lower cost per token.

Same Quality, Faster

Models with similar or better performance but lower latency.

Same Cost, Better

Models at a similar price point with higher benchmark scores.

No alternatives found

Other Models from Z.AI

Compare performance with other models from the same creator

ModelScoreLatencyCost/1M
Z.AI: GLM 4.592.1%48.4s$0.95
Z.AI: GLM 4.5 Air86.5%35.6s$0.14
Z.AI: GLM 4.5 Air85.6%90.8sFree
Z.AI: GLM 4.782.7%56.9s$0.95
Z.AI: GLM 4.5V80.8%51.3s$1.20
Z.AI: GLM 4.6V71.4%57.3s$0.60
Z.AI: GLM 4 32B 67.7%18.7s$0.10
Z.AI: GLM 4.7 Flash54.2%175.4s$0.23

Benchmark Performance

How this model performs across different benchmarks

No benchmark data available

Run benchmarks with this model to see performance breakdown

Price vs Performance

Compare cost efficiency across all models

Current model (baseline)
Other models (relative score)
Y-axis shows score difference from shared benchmarks. X-axis uses log scale.

Score Over Time

Performance trends across all benchmark runs

Benchmark Activity

Number of benchmark runs over time

Quickstart

Get started with this model using OpenRouter

View on OpenRouter
import { OpenRouter } from "@openrouter/sdk";

const openrouter = new OpenRouter({
  apiKey: "<OPENROUTER_API_KEY>"
});

const completion = await openrouter.chat.completions.create({
  model: "z-ai/glm-4.6:exacto",
  messages: [
    {
      role: "user",
      content: "Hello!"
    }
  ]
});

console.log(completion.choices[0].message.content);

Get your API key at openrouter.ai/keys