EN DIRECT
Apr 7Claude Mythos Preview · Anthropic's most capable model arrives·Mar 31GPT-5.4 Nano launched on OpenAI·Mar 31GPT-5.4 Mini joins the OpenAI lineup·Mar 30Claude Opus 4.5 input price dropped · $5.00 per 1M tokens·Mar 30Mistral Small 4 available via Mistral AI·Mar 29Gemini 2.5 Pro scores 94.1% on MMLU·Mar 29Grok 4.20 Multi-Agent Beta enters agent rankings·Mar 28DeepSeek V3.2 output price dropped · $0.38 per 1M tokens·Mar 287 new MCP servers added in dev-tools category·Mar 27Claude Sonnet 4.6 released by Anthropic·Mar 27Claude Opus 4.6 released by Anthropic·Mar 26OTIS Mock AIME 2024-2025 benchmark added·Mar 26Claude Opus 4.1 pricing increased · $15/$75 per 1M tokens·Mar 25Grok 4.20 Beta launched by xAI·Mar 25Inception added as a tracked provider·Mar 24DeepSeek R1 0528 posted 87.2% on GPQA Diamond·Mar 243 new MCP servers in AI/ML category·Mar 23GPT-4o Audio Preview marked as deprecated·Mar 23Mistral Medium 3.1 input price cut to $0.40 per 1M tokens·Mar 22DeepSeek V3.2 Speciale released·Mar 22WeirdML benchmark now tracked on BenchGecko·Mar 20Nemotron 3 Super (120B) launched by NVIDIA·Mar 20Gemini 2.5 Flash Lite priced at $0.10/$0.40 per 1M tokens·Mar 18Mistral Large 3 2512 released by Mistral AI·Mar 18Grok Code Fast 1 added to agent rankings·Mar 16Claude Sonnet 4.5 scores 91.7% on MMLU·Mar 1612 new MCP servers added across 5 categories·Mar 14GPT-5.4 Pro launched · OpenAI's new flagship·Mar 14GPT-5.4 standard tier released by OpenAI·Mar 12Grok 3 Mini marked as deprecated by xAI·Mar 12Llama 3.3 Nemotron Super 49B pricing dropped·Mar 10Liquid added as a tracked provider·Mar 10MiniMax M2.7 released by MiniMax·Mar 8Grok 4 posted 89.4% on GPQA Diamond·Mar 8LAMBADA benchmark scores now tracked·Mar 5Gemini 2.5 Flash output price reduced to $2.50 per 1M tokens·Mar 5Mercury 2 launched by Inception·Mar 3Qwen3.5-Flash released by Alibaba Qwen·Mar 35 new MCP servers added · finance and auth categories·Apr 7Claude Mythos Preview · Anthropic's most capable model arrives·Mar 31GPT-5.4 Nano launched on OpenAI·Mar 31GPT-5.4 Mini joins the OpenAI lineup·Mar 30Claude Opus 4.5 input price dropped · $5.00 per 1M tokens·Mar 30Mistral Small 4 available via Mistral AI·Mar 29Gemini 2.5 Pro scores 94.1% on MMLU·Mar 29Grok 4.20 Multi-Agent Beta enters agent rankings·Mar 28DeepSeek V3.2 output price dropped · $0.38 per 1M tokens·Mar 287 new MCP servers added in dev-tools category·Mar 27Claude Sonnet 4.6 released by Anthropic·Mar 27Claude Opus 4.6 released by Anthropic·Mar 26OTIS Mock AIME 2024-2025 benchmark added·Mar 26Claude Opus 4.1 pricing increased · $15/$75 per 1M tokens·Mar 25Grok 4.20 Beta launched by xAI·Mar 25Inception added as a tracked provider·Mar 24DeepSeek R1 0528 posted 87.2% on GPQA Diamond·Mar 243 new MCP servers in AI/ML category·Mar 23GPT-4o Audio Preview marked as deprecated·Mar 23Mistral Medium 3.1 input price cut to $0.40 per 1M tokens·Mar 22DeepSeek V3.2 Speciale released·Mar 22WeirdML benchmark now tracked on BenchGecko·Mar 20Nemotron 3 Super (120B) launched by NVIDIA·Mar 20Gemini 2.5 Flash Lite priced at $0.10/$0.40 per 1M tokens·Mar 18Mistral Large 3 2512 released by Mistral AI·Mar 18Grok Code Fast 1 added to agent rankings·Mar 16Claude Sonnet 4.5 scores 91.7% on MMLU·Mar 1612 new MCP servers added across 5 categories·Mar 14GPT-5.4 Pro launched · OpenAI's new flagship·Mar 14GPT-5.4 standard tier released by OpenAI·Mar 12Grok 3 Mini marked as deprecated by xAI·Mar 12Llama 3.3 Nemotron Super 49B pricing dropped·Mar 10Liquid added as a tracked provider·Mar 10MiniMax M2.7 released by MiniMax·Mar 8Grok 4 posted 89.4% on GPQA Diamond·Mar 8LAMBADA benchmark scores now tracked·Mar 5Gemini 2.5 Flash output price reduced to $2.50 per 1M tokens·Mar 5Mercury 2 launched by Inception·Mar 3Qwen3.5-Flash released by Alibaba Qwen·Mar 35 new MCP servers added · finance and auth categories·

L'économie de l'IA, Suivie.

Pouls20·sain
Bulle278%·agité
GPT-5.5 Pro+4.0
Open Source16.2%

Le Pouls

Gecko Pouls
healthy
7d · +3 pts
Indice Bulle · composantes
Valuation Premiumhealthy+2.1
Funding Accelerationhealthy+1.5
Concentration Riskhealthy0
Revenue Qualityhealthy+1.4
Capex Gaphealthy+0.3
Plus forte variation · Valuation Premium hausse 2.1 pts
Indice Bulle IA
SainMousseuxSurchaufféBulle
Mis à jour May 14·Méthodologie·Recherche·API gratuite·Développeurs

Signaux transversaux

La Matrice complète
#Benchmarks
1OpenAI logoGPT-5.5 Pro99.9$30.00400K3
2Anthropic logoClaude Mythos Preview99.81000K14
3Alibaba Qwen logoQwen3.5 397B A17B96.3$0.39262K11
4DeepSeek logoDeepSeek V3.2 Speciale95.2$0.40164K9
5OpenAI logoGPT-5.4 Pro93.0$30.001050K8
6OpenAI logoGPT-5.1-Codex-Max91.2$1.25400K8
7Google DeepMind logoGemini 3.1 Pro Preview90.0$2.001049K23
8stepfun logoStep 3.5 Flash89.5$0.10262K10
9OpenAI logoGPT-5 Chat89.0$1.25128K7
10Alibaba Qwen logoQwen3.6 Plus88.7$0.331000K11
11DeepSeek logoDeepSeek R1 Distill Qwen 14B88.311
12
HA
Qwen2.5 72B Instruct Abliterated
87.56
13z-ai logoGLM 5.187.0$1.05203K12
14OpenAI logoGPT-5.2-Codex85.4$1.75400K9
15Anthropic logoClaude Instant84.64
16DeepSeek logoDeepSeek-V2 (MoE-236B, May 2024)84.47
17OpenAI logoGPT-5.483.4$2.501050K16
18Anthropic logoClaude Opus 4.6 (Fast)83.3$30.001000K12
19OpenAI logoGPT-5.1-Codex82.8$1.25400K8
20xiaomi logoMiMo-V2-Flash81.7$0.09262K11
Tous les parcours
Rechercher 297 termes IA · des transformers à l'attention premiumOuvrir
Méthodologie complète
À quelle fréquence les données BenchGecko sont-elles mises à jour ?

Les données de modèles et benchmarks sont actualisées quotidiennement depuis les sources primaires. Les tarifs sont collectés auprès de chaque API fournisseur en continu. Les signaux de notoriété sont agrégés chaque semaine. Le Gecko Pouls est recalculé à 00:00 UTC.

Qu'est-ce que le Gecko Pouls ?

Un score composite de 0 à 100 mesurant la santé de l'économie IA. Il combine l'inverse de l'Indice Bulle, la vélocité des benchmarks, la compression des prix, la diversité de la notoriété et la tension de la chaîne d'approvisionnement en un seul chiffre. Plus bas = plus sain.

Comment les scores de benchmark sont-ils normalisés ?

Chaque benchmark est normalisé min-max sur l'ensemble des modèles évalués. Les classements font la moyenne des scores normalisés sur 3 benchmarks ou plus par modèle pour éviter de surpondérer un seul test.

D'où viennent les données de tarification ?

Directement des API fournisseurs · OpenRouter, OpenAI, Anthropic, Google, xAI, DeepSeek, Mistral et d'autres. Chaque capture est mise en cache avec attribution de la source sur la page détail du modèle.

Puis-je citer les données BenchGecko ?

Oui. Chaque page dispose d'une barre Partager et Citer avec les formats APA, MLA, BibTeX, Chicago et texte brut. L'attribution est requise sur le tier API gratuit et encouragée partout.

Sources ·OpenRouterEpoch AISWE-benchMCP RegistryChatbot ArenaHuggingFaceLiveBenchArtificial AnalysisSEALAider
Mis à jour il y a 2h · 10+ sources fiables · zéro contenu éditorial·Learn · Glossary·Recherche·Developers