LIVETracking 994 AI models from 267 providers.
BenchGeckoBeta
AI
Providers
Economy
Mindshare
Infra
Learn
Labs
Developers
Gecko Tests·powered by GeckoBench · AI Bias, Censorship, IQ & PoliticsView Gecko TestsBuild your own chart
Home·Models
BenchGecko

BenchGecko is the data layer of the AI economy. Thousands of models with cross-provider pricing and daily price history. Hundreds of companies with valuations, funding timelines, and revenue estimates. Benchmark scores, developer adoption signals, agent leaderboards, and a changelog that captures every price drop, every launch, every deprecation as it happens. If it moved in AI today, it's already on BenchGecko.

@BenchGecko

Rankings

  • Leaderboard
  • All Models
  • AI Agents
  • Benchmarks
  • Compare Models
  • MCP Servers
  • Skills
  • Pricing

Top Models

  • Qwen3 VL 235B A22B Instruct
  • DeepSeek V3.1 Terminus
  • Qwen3 VL 235B A22B Thinking
  • Trinity Large Preview (free)
  • GLM 4.5 Air
  • GLM 4.7 Flash
  • MiniMax M1
  • GLM 4.5V
  • Gemma 3 12B
  • Nova 2 Lite

Providers

  • AI Research
  • AI21 Labs
  • AI4Bharat
  • aion-labs
  • alfredpros
  • Alibaba
  • Alimama Creative
  • Allen Dou
  • All Providers →

Benchmarks

  • Aider · Code Editing
  • Aider polyglot
  • ANLI
  • APEX-Agents
  • ARC AI2
  • ARC-AGI
  • ARC-AGI-2
  • Artificial Analysis · Agentic Index
  • All Benchmarks →

Resources

  • API Documentation
  • Methodology
  • Research
  • Gecko Mindshare
  • News
  • Changelog
  • Countries
  • Model Status
  • Brand Assets
  • Developers
  • Charts
  • Learn
  • Economy
  • Labs
  • Gecko Tests
© 2026 BenchGecko·The AI Economy, Tracked
Privacy·Terms·Press
Models · LeaderboardUpdated 8d ago · 994 models · 267 providers · 128 benchmarks

Every AI Model · Tracked

The most complete list of AI models you can actually use · 994 models, 267 providers, 128 benchmarks · all scored, priced, and ranked in one place.

Benchmark scoredCross-provider pricingContext · parameters · licensingRefreshed daily
Compare side by sideShowing 47 of 994
Total Models
994
Providers
267
Top Model
GPT-5.5 Pro
87.8%
Cheapest Capable
gpt-oss-20b
$0.03/M
PodiumMoversLeaderboardFAQ

Top 10 Overall

Ranked by average benchmark score · min 3 benchmarks

#1 GOLD3 benchmarks
OpenAI logo
GPT-5.5 Pro
OpenAI
87.8%
#2 SILVER6 benchmarks
OpenAI logo
GPT-5.5
OpenAI
85.0%
#3 BRONZE7 benchmarks
OpenAI logo
GPT-5 Chat
OpenAI
81.9%
#4Anthropic logo
Claude Mythos Preview
Anthropic
81.8%14 benchmarks
#5Alibaba Qwen logo
Qwen3.5 397B A17B
Alibaba Qwen
78.4%11 benchmarks
#6DeepSeek logo
DeepSeek V3.2 Speciale
DeepSeek
78.2%9 benchmarks
#7Anthropic logo
Claude Instant
Anthropic
78.0%4 benchmarks
#8stepfun logo
Step 3.5 Flash
stepfun
76.9%10 benchmarks
#9DeepSeek logo
DeepSeek-V2 (MoE-236B, May 2024)
DeepSeek
76.5%7 benchmarks
#10xiaomi logo
MiMo-V2-Flash
xiaomi
73.3%11 benchmarks

What's moving

New releases, coverage, price leaders, and ELO champions

#ModelProviderCategoryReleased
1ibm-granite logoGranite 4.1 8Bibm-graniteLLMApr 30, 2026
2xAI logoGrok 4.3xAIMultimodalApr 30, 2026
3NVIDIA logoNemotron 3 Nano Omni (free)NVIDIAMultimodalApr 28, 2026
4openrouter logoOwl AlphaopenrouterLLMApr 28, 2026
5Alibaba Qwen logoQwen3.5 Plus 2026-04-20Alibaba QwenMultimodalApr 27, 2026
6Alibaba Qwen logoQwen3.6 27BAlibaba QwenMultimodalApr 27, 2026
7Alibaba Qwen logoQwen3.6 35B A3BAlibaba QwenMultimodalApr 27, 2026
8Alibaba Qwen logoQwen3.6 FlashAlibaba QwenMultimodalApr 27, 2026
9Alibaba Qwen logoQwen3.6 Max PreviewAlibaba QwenLLMApr 27, 2026
10DeepSeek logoDeepSeek V4 FlashDeepSeekLLMApr 24, 2026
Provider
Filter
47 results

All Models

#ModelProviderCategoryContextIn $/MOut $/MAvgBenchmarksELO
1Meta logoLlama 3.3 70B InstructOSSMetaLLM131K$0.10$0.3246.9%8—
2Meta logoMeta Llama 3 8B InstructOSSMetaLLM—TBDTBD45.2%25—
3Meta logoMeta Llama 3 8BOSSMetaLLM—TBDTBD44.2%11—
4Meta logoLlama 2-13BOSSMetaLLM—TBDTBD42.5%14—
5Meta logoLlama 3.1 405BOSSMetaLLM—TBDTBD38.0%21—
6Meta logoLlama 3.1 70B InstructOSSMetaLLM131K$0.40$0.4037.8%16—
7Meta logoLlama 3.2 90BOSSMetaLLM—TBDTBD36.1%6—
8Meta logoLLaMA-13BOSSMetaLLM—TBDTBD34.9%20—
9Meta logoLlama 3 70B InstructOSSMetaLLM8K$0.51$0.7432.4%9—
10Meta logoLlama 3 8B InstructOSSMetaLLM8K$0.03$0.0430.8%16—
11Meta logoLlama 3.3 70B Instruct (free)OSSMetaLLM66KFreeFree29.1%8—
12Meta logoLlama 4 MaverickOSSMetaMultimodal1.0M$0.15$0.6028.0%17—
13Meta logoLlama 3.1 8B InstructOSSMetaLLM16K$0.02$0.0527.4%16—
14Meta logoLlama 2 7b Chat HfOSSMetaLLM—TBDTBD27.3%11—
15Meta logoLlama 3.2 3B InstructOSSMetaLLM80K$0.05$0.3424.2%7—
16Meta logoLlama 2 7b HfOSSMetaLLM—TBDTBD23.6%11—
17Meta logoLlama 4 ScoutOSSMetaMultimodal328K$0.08$0.3018.9%11—
18Meta logoLlama 3.2 1B InstructOSSMetaLLM60K$0.03$0.2014.5%7—
19Meta logoLlama 3.2 3B Instruct (free)OSSMetaLLM131KFreeFree8.7%6—
20Meta logoHf Seamless M4t MediumOSSMetaLLM—TBDTBD—0—
21Meta logoHubert Large Ls960 FtOSSMetaLLM—TBDTBD—0—
22Meta logoLlama 3.2 11B Vision InstructOSSMetaMultimodal131K$0.24$0.24—0—
23Meta logoLlama 4 Scout 17B 16E InstructMetaLLM—TBDTBD—1—
24Meta logoLlama Guard 3 8BOSSMetaLLM131K$0.48$0.03—0—
25Meta logoLlama Guard 4 12BOSSMetaMultimodal164K$0.18$0.18—0—
26Meta logoLlama Guard 4 12B (free)MetaMultimodal164KFreeFree—0—
27Meta logoMms 1b AllOSSMetaLLM—TBDTBD—0—
28Meta logoMms Tts EngOSSMetaLLM—TBDTBD—0—
29Meta logoMms Tts HatOSSMetaLLM—TBDTBD—0—
30Meta logoMms Tts HinOSSMetaLLM—TBDTBD—0—
31Meta logoMms Tts KikOSSMetaLLM—TBDTBD—0—
32Meta logoMms Tts KinOSSMetaLLM—TBDTBD—0—
33Meta logoMms Tts KorOSSMetaLLM—TBDTBD—0—
34Meta logoMms Tts OrmOSSMetaLLM—TBDTBD—0—
35Meta logoMms Tts RusOSSMetaLLM—TBDTBD—0—
36Meta logoMms Tts SwhOSSMetaLLM—TBDTBD—0—
37Meta logoNougat BaseOSSMetaLLM—TBDTBD—0—
38Meta logoOpt 125mMetaLLM—TBDTBD—0—
39Meta logoRoberta BaseOSSMetaLLM—TBDTBD—0—
40Meta logoRoberta LargeOSSMetaLLM—TBDTBD—0—
41Meta logoS2t Small Librispeech AsrOSSMetaLLM—TBDTBD—0—
42Meta logoSeamless M4t V2 LargeOSSMetaLLM—TBDTBD—0—
43Meta logoWav2vec2 Base 960hOSSMetaLLM—TBDTBD—0—
44Meta logoWav2vec2 Conformer Rope Large 960h FtOSSMetaLLM—TBDTBD—0—
45Meta logoWav2vec2 Xlsr 53 Espeak Cv FtOSSMetaLLM—TBDTBD—0—
46Meta logoXlm Roberta BaseOSSMetaLLM—TBDTBD—0—
47Meta logoXlm Roberta LargeOSSMetaLLM—TBDTBD—0—

47 rows · click column headers to sort · pick up to 4 models to compare

Top 5 by category

Specialist leaders across every modality we track

LLMs

See all →
#1OpenAI logoGPT-5.5 Pro87.8%#2OpenAI logoGPT-5.585.0%#3Anthropic logoClaude Mythos Preview81.8%#4DeepSeek logoDeepSeek V3.2 Speciale78.2%#5Anthropic logoClaude Instant78.0%

Multimodal

See all →
#1OpenAI logoGPT-5 Chat81.9%#2Alibaba Qwen logoQwen3.5 397B A17B78.4%#3Google DeepMind logoGemini 2.5 Pro Preview 05-0676.9%#4OpenAI logoGPT-5.1-Codex-Max72.0%#5OpenAI logoo4 Mini High72.0%

Decision shortcuts

Jump from the leaderboard into comparison, provider, pricing, and benchmark paths

GPT-5.5 Pro vs Qwen3.5
Frontier comparison
Qwen3.5 vs DeepSeek V3.2
Open model comparison
OpenAI model lineup
Provider profile
Anthropic Claude lineup
Provider profile
Coding model pricing
Cost by use case
Reasoning model pricing
Cost by use case
SWE-bench Verified results
Coding benchmark
GPQA Diamond results
Reasoning benchmark

Frequently asked

Quick answers, sourced from our data

How many AI models does BenchGecko track?

BenchGecko currently tracks 994 AI models across 267 providers, each scored against up to 128 benchmarks. New models are added continuously and the full dataset refreshes daily.

What is the best AI model right now?

"Best" depends on the task. For general reasoning we rank by average score across 3+ benchmarks; for coding we surface ELO and SWE-bench specifically; for cost/performance we expose a "cheapest capable" metric. Use the filter bar and column sort to define your own winner, or pick a category from the mini tables below.

How is the average score calculated?

Average score is the arithmetic mean of a model's normalized benchmark scores, computed only when the model has at least one public benchmark result. Models with fewer than 3 benchmarks are excluded from the podium to avoid single-score outliers.

Where does BenchGecko get this data?

Model metadata and pricing come from OpenRouter's public API. Benchmarks are pulled from Epoch AI (CC-BY) and SWE-bench's public leaderboards. ELO ratings come from LMArena. Everything is re-normalized and cross-linked daily. See the methodology page for full provenance.

Can I use this data in my article or product?

Yes. All BenchGecko data is licensed CC BY 4.0 — attribution required. Use the "Cite this page" button above for ready-made APA, MLA, AP Style, BibTeX, and HTML embed snippets. The free API tier requires a backlink to benchgecko.ai.

How often does the data refresh?

Core model, pricing and benchmark data refreshes every 24 hours. Live status and pricing alerts can fire more frequently when upstream sources change. The "Live" pill on this page lights up when the last refresh was less than an hour ago.

Which models are open source?

Toggle the "Open Source" filter to see only models with OSS weights available. We currently count permissively — weights-available licenses like Llama, Qwen, DeepSeek, and Mistral Open count as OSS for this filter even when the weights come with usage restrictions.

How do I compare two models side by side?

Check the boxes next to any 2-4 models in the leaderboard above, then click "Compare →" in the pink action bar. You can also navigate directly to /compare/[modelA]-vs-[modelB] for shareable comparisons.

See also

Related model, benchmark, provider, and pricing indexes

Compare models
Side-by-side with charts
Benchmarks
Every eval we track
Providers
Where to run models
Pricing
Every model, every provider
AI agents
SWE-bench leaderboard
MCP servers
Tools for LLMs
AI economy
Valuations & funding
Methodology
How we score