All models

Qwen: Qwen3 Coder 480B A35B

by Qwen

Qwen3-Coder-480B-A35B-Instruct is a Mixture-of-Experts (MoE) code generation model developed by the Qwen team. It is optimized for agentic coding tasks such as function calling, tool use, and long-context reasoning over repositories. The model features 480 billion total parameters, with 35 billion active per forward pass (8 out of 160 experts). Pricing for the Alibaba endpoints varies by context length. Once a request is greater than 128k input tokens, the higher pricing is used.

Avg Score

83.4%

22 answers

Avg Latency

13.6s

9 runs

Pricing

Free

input

/

Free

output

per 1M tokens

Context

262K

tokens

Alternatives

Models with similar or better quality but different tradeoffs

Same Quality, Cheaper

Models with similar or better performance at a lower cost per token.

Same Quality, Faster

Models with similar or better performance but lower latency.

Same Cost, Better

Models at a similar price point with higher benchmark scores.

Other Models from Qwen

Compare performance with other models from the same creator

ModelScoreLatencyCost/1M
Qwen: Qwen3 VL 235B A22B Instruct88.6%48.0s$0.70
Qwen: Qwen3 Max87.9%31.7s$3.60
Qwen: Qwen-Plus85.8%22.0s$0.80
Qwen: Qwen3 235B A22B Thinking 250784.6%248.4s$0.35
Qwen: Qwen3 Coder Plus84.2%20.5s$3.00
Qwen: Qwen3 Next 80B A3B Thinking80.8%28.8s$0.68
Qwen: Qwen3 VL 235B A22B Thinking80.4%112.0s$1.98
Qwen: Qwen3 VL 32B Instruct78.8%26.3s$1.00
Qwen: Qwen Plus 072878.5%19.6s$0.80
Qwen: Qwen3 235B A22B74.6%78.3s$0.40
Qwen: Qwen3 30B A3B Instruct 250773.1%30.8s$0.20
Qwen: Qwen3 Next 80B A3B Instruct71.4%27.9s$0.59
Qwen: Qwen3 VL 8B Thinking70.8%79.4s$1.14
Qwen: Qwen-Turbo70.7%18.6s$0.13
Qwen: Qwen Plus 072870.0%49.2s$2.20
Qwen: Qwen3 235B A22B Instruct 250769.6%25.6s$0.27
Qwen: Qwen3 Coder Flash69.2%14.1s$0.90
Qwen: Qwen3 30B A3B Thinking 250765.8%45.6s$0.20
Qwen: Qwen-Max 65.8%12.7s$4.00
Qwen: Qwen2.5 VL 72B Instruct65.0%22.1s$0.38
Qwen: Qwen3 Coder 30B A3B Instruct63.6%30.8s$0.17
Qwen: Qwen3 VL 30B A3B Thinking63.1%83.6s$0.60
Qwen2.5 72B Instruct62.9%22.2s$0.26
Qwen: Qwen3 8B62.5%111.5s$0.15
Qwen: Qwen3 14B61.5%70.4s$0.14
Qwen: Qwen VL Max61.3%36.4s$2.00
Qwen: Qwen3 32B59.6%121.7s$0.16
Qwen: QwQ 32B59.2%143.6s$0.28
Qwen: Qwen VL Plus57.9%10.4s$0.42
Qwen: Qwen3 30B A3B57.7%149.8s$0.14
Qwen: Qwen3 VL 30B A3B Instruct57.5%38.3s$0.38
Qwen: Qwen3 VL 8B Instruct54.6%133.7s$0.29
Qwen2.5 Coder 32B Instruct52.7%19.7s$0.07
Qwen: Qwen2.5 VL 32B Instruct47.3%39.1s$0.14
Qwen: Qwen2.5-VL 7B Instruct31.3%20.7sFree
Qwen: Qwen2.5 7B Instruct30.8%12.6s$0.07
Qwen: Qwen2.5-VL 7B Instruct21.7%10.9s$0.20
Qwen: Qwen2.5 Coder 7B Instruct6.7%6.4s$0.06
Qwen: Qwen3 Next 80B A3B InstructFree
Qwen: Qwen3 4B (free)Free

Benchmark Performance

How this model performs across different benchmarks

No benchmark data available

Run benchmarks with this model to see performance breakdown

Price vs Performance

Compare cost efficiency across all models

Current model (baseline)
Other models (relative score)
Y-axis shows score difference from shared benchmarks. X-axis uses log scale.

Score Over Time

Performance trends across all benchmark runs

Benchmark Activity

Number of benchmark runs over time

Quickstart

Get started with this model using OpenRouter

View on OpenRouter
import { OpenRouter } from "@openrouter/sdk";

const openrouter = new OpenRouter({
  apiKey: "<OPENROUTER_API_KEY>"
});

const completion = await openrouter.chat.completions.create({
  model: "qwen/qwen3-coder:free",
  messages: [
    {
      role: "user",
      content: "Hello!"
    }
  ]
});

console.log(completion.choices[0].message.content);

Get your API key at openrouter.ai/keys