ベータ
ランキング/Gemini 2.5 Pro
Google DeepMind logo

Gemini 2.5 Pro

開発元 Google DeepMind · リリース日 2025-06-17

56.2
平均スコア
$1.25/1M
入力料金
$10.00/1M
出力料金
1.0M tokens (~524 books)
コンテキストウィンドウ
multimodal
タイプ

Tested on 42 benchmarks with 56.2% average. Top scores: Chatbot Arena Elo — Overall (1448.2%), Chatbot Arena Elo — Coding (1202.0%), MATH level 5 (95.6%).

ベンチマークカテゴリスコアBar
Chatbot Arena Elo — Overallarena1448.2
Chatbot Arena Elo — Codingarena1202.0
MATH level 5math95.6
Fiction.LiveBenchknowledge91.7
OpenCompass — IFEvallanguage90.0
OpenCompass — AIME2025math88.7
HELM — MMLU-Proknowledge86.3
Lech Mazur Writingknowledge86.0
OpenCompass — MMLU-Proknowledge85.8
HELM — WildBenchreasoning85.7
OTIS Mock AIME 2024-2025math84.7
OpenCompass — GPQA-Diamondknowledge84.7
HELM — IFEvallanguage84.0
Aider polyglotcoding83.1
GeoBenchknowledge81.0
GPQA diamondknowledge80.4
HELM — GPQAknowledge74.9
OpenCompass — LiveCodeBenchV6coding71.3
CadEvalcoding64.0
SWE-Bench verifiedcoding57.6
SimpleQA Verifiedknowledge56.0
SimpleBenchreasoning54.9
WeirdMLcoding54.0
DeepResearch Benchknowledge49.7
AudioMultiChallengeknowledge46.9
AudioMultiChallenge — Text Outputknowledge46.9
Balrogknowledge43.3
HELM — Omni-MATHmath41.6
ARC-AGIreasoning41.0
Artificial Analysis — Quality Indexspeed34.6
Artificial Analysis — Agentic Indexspeed32.7
Terminal Benchcoding32.6
Artificial Analysis — Coding Indexspeed31.9
The Agent Companyagentic30.3
OpenCompass — HLEknowledge21.1
Chess Puzzlesknowledge20.0
VPCTknowledge19.6
HLEknowledge17.7
FrontierMath-2025-02-28-Privatemath14.1
ARC-AGI-2reasoning4.9
FrontierMath-Tier-4-2025-07-01-Privatemath4.2
GSO-Benchcoding3.9