Home/Models/Llama 2-13B
Meta logo

Llama 2-13B

by Meta · Released Jan 2024

Open Source
40.7
avg score
Rank #142
Compare
Better than 39% of all models
Context
N/A
Input $/1M
TBD
Output $/1M
TBD
Type
text
License
Open Source
Benchmarks
14 tested
Data updated today
About

Tested on 14 benchmarks with 42.5% average. Top scores: TriviaQA (79.6%), LAMBADA (76.5%), HellaSwag (74.3%).

Capabilities
reasoning
44.3
#58 globally
math
20.1
#159 globally
knowledge
46.5
#116 globally
Benchmark Scores
Compare All
Tested on 14 benchmarks · Ranked across 3 categories
Score Distribution (all 233 models)
0255075100
▲ You are here
BBH

BIG-Bench Hard. 23 challenging tasks from BIG-Bench where prior language models fell below average human performance.

44.3
GSM8K

Grade school math word problems. 8,500 problems testing multi-step arithmetic reasoning. A foundational math benchmark.

36.9
MATH level 5

Competition-level math from AMC, AIME, and olympiad problems. Level 5 is the hardest tier, requiring creative problem-solving.

3.3
TriviaQA

Trivia questions sourced from trivia enthusiasts and quiz websites. Tests breadth of general knowledge.

79.6
LAMBADA

Language modeling benchmark testing ability to predict the last word of passages requiring long-range context understanding.

76.5
HellaSwag

Sentence completion requiring commonsense reasoning about physical and social situations. Tests real-world understanding.

74.3
Excellent (85+) Good (70-85) Average (50-70) Below (<50)
Links
Documentation
Community
BenchGecko API
llama-2-13b
Specifications
  • Typetext
  • ContextN/A
  • ReleasedJan 2024
  • LicenseOpen Source
  • Statusbenchmark-only
Available On
Meta logoMetaTBD
Share & Export
Tweet
Llama 2-13B is an open-source text AI model by Meta, released in January 2024. It has an average benchmark score of 40.7.