Home/Models/GPT-4o (2024-08-06)
OpenAI logo

GPT-4o (2024-08-06)

by OpenAI · Released Aug 2024

Multimodal
39.5
avg score
Rank #147
Compare
Better than 37% of all models
Context
128K tokens (~64 books)
Input $/1M
$2.50
Output $/1M
$10.00
Type
multimodal
License
Proprietary
Benchmarks
11 tested
Data updated today
About

The 2024-08-06 version of GPT-4o offers improved performance in structured outputs, with the ability to supply a JSON schema in the respone_format. Read more [here](https://openai.com/index/introducing-structured-outputs-in-the-api/). GPT-4o ("o" for "omni") is...

Tested on 11 benchmarks with 35.6% average. Top scores: Chatbot Arena Elo — Overall (1334.3%), MMLU (79.1%), Aider — Code Editing (71.4%).

Looking for similar performance at lower cost?
ERNIE 4.5 21B A3B Thinking scores 39.8 (101% as good) at $0.07/1M input · 97% cheaper
Capabilities
coding
40.2
#96 globally
reasoning
1.4
#182 globally
math
20.0
#161 globally
knowledge
55.7
#74 globally
multimodal
62.5
#3 globally
Benchmark Scores
Compare All
Tested on 11 benchmarks · Ranked across 6 categories
Score Distribution (all 233 models)
0255075100
▲ You are here
Aider — Code Editing

Code editing benchmark from the Aider project. Measures ability to apply targeted code changes while maintaining correctness and style.

71.4
CadEval

Computer-aided design evaluation. Tests understanding of CAD concepts, 3D modeling, and engineering design principles.

26.0
Aider polyglot

Multi-language code editing from Aider. Tests editing ability across Python, JavaScript, TypeScript, Java, C++, Go, Rust, and more.

23.1
SimpleBench

Deceptively simple questions that humans find easy but AI models often get wrong. Tests common sense and reasoning gaps.

1.4
MATH level 5

Competition-level math from AMC, AIME, and olympiad problems. Level 5 is the hardest tier, requiring creative problem-solving.

53.3
OTIS Mock AIME 2024-2025

Mock AIME (American Invitational Mathematics Exam) problems from OTIS. Tests mathematical competition performance.

6.3
FrontierMath-2025-02-28-Private

Original research-level math problems created by professional mathematicians. Problems are unpublished and cannot be memorized.

0.3
Excellent (85+) Good (70-85) Average (50-70) Below (<50)
Links
Documentation
Community
BenchGecko API
gpt-4o-2024-08-06
Specifications
  • Typemultimodal
  • Context128K tokens (~64 books)
  • ReleasedAug 2024
  • LicenseProprietary
  • StatusActive
  • Cost / Message~$0.015
Available On
OpenAI logoOpenAI$2.50
Share & Export
Tweet
GPT-4o (2024-08-06) is a proprietary multimodal AI model by OpenAI, released in August 2024. It has an average benchmark score of 39.5. Context window: 128K tokens.