Compare · ModelsLive · 2 picked · head to head
Gemini 2.5 Pro vs Qwen3 32B
Side by side · benchmarks, pricing, and signals you can act on.
Winner summary
Gemini 2.5 Pro wins on 8/8 benchmarks
Gemini 2.5 Pro wins 8 of 8 shared benchmarks. Leads in coding · arena · math.
Category leads
coding·Gemini 2.5 Proarena·Gemini 2.5 Promath·Gemini 2.5 Proknowledge·Gemini 2.5 Prolanguage·Gemini 2.5 Pro
Hype vs Reality
Attention vs performance
Gemini 2.5 Pro
#59 by perf·no signal
Qwen3 32B
#48 by perf·no signal
Best value
Qwen3 32B
36.4x better value than Gemini 2.5 Pro
Gemini 2.5 Pro
10.0 pts/$
$5.63/M
Qwen3 32B
363.8 pts/$
$0.16/M
Vendor risk
Who is behind the model
Google DeepMind
$4.00T·Tier 1
Alibaba (Qwen)
$293.0B·Tier 1
Head to head
8 benchmarks · 2 models
Gemini 2.5 ProQwen3 32B
Aider polyglot
Gemini 2.5 Pro leads by +43.1
Aider Polyglot · measures how well AI models can edit code across multiple programming languages using the Aider coding assistant framework.
Gemini 2.5 Pro
83.1
Qwen3 32B
40.0
Chatbot Arena Elo · Overall
Gemini 2.5 Pro leads by +101.2
Gemini 2.5 Pro
1448.2
Qwen3 32B
1347.0
OpenCompass · AIME2025
Gemini 2.5 Pro leads by +18.4
Gemini 2.5 Pro
88.7
Qwen3 32B
70.3
OpenCompass · GPQA-Diamond
Gemini 2.5 Pro leads by +17.4
Gemini 2.5 Pro
84.7
Qwen3 32B
67.3
OpenCompass · HLE
Gemini 2.5 Pro leads by +12.6
Gemini 2.5 Pro
21.1
Qwen3 32B
8.5
OpenCompass · IFEval
Gemini 2.5 Pro leads by +4.0
Gemini 2.5 Pro
90.0
Qwen3 32B
86.0
OpenCompass · LiveCodeBenchV6
Gemini 2.5 Pro leads by +13.7
Gemini 2.5 Pro
71.3
Qwen3 32B
57.6
OpenCompass · MMLU-Pro
Gemini 2.5 Pro leads by +7.8
Gemini 2.5 Pro
85.8
Qwen3 32B
78.0
Full benchmark table
| Benchmark | Gemini 2.5 Pro | Qwen3 32B |
|---|---|---|
Aider polyglot Aider Polyglot · measures how well AI models can edit code across multiple programming languages using the Aider coding assistant framework. | 83.1 | 40.0 |
Chatbot Arena Elo · Overall | 1448.2 | 1347.0 |
OpenCompass · AIME2025 | 88.7 | 70.3 |
OpenCompass · GPQA-Diamond | 84.7 | 67.3 |
OpenCompass · HLE | 21.1 | 8.5 |
OpenCompass · IFEval | 90.0 | 86.0 |
OpenCompass · LiveCodeBenchV6 | 71.3 | 57.6 |
OpenCompass · MMLU-Pro | 85.8 | 78.0 |
Pricing · per 1M tokens · projected $/mo at 10M tokens
| Model | Input | Output | Context | Projected $/mo |
|---|---|---|---|---|
| $1.25 | $10.00 | 1.0M tokens (~524 books) | $34.38 | |
| $0.08 | $0.24 | 41K tokens (~20 books) | $1.20 |