Tested on 17 benchmarks with 44.4% average. Top scores: HellaSwag (85.3%), TriviaQA (79.9%), LAMBADA (79.8%).
BIG-Bench Hard. 23 challenging tasks from BIG-Bench where prior language models fell below average human performance.
HuggingFace MuSR (Multi-Step Reasoning). Tests multi-hop reasoning requiring chaining multiple facts together.
Grade school math word problems. 8,500 problems testing multi-step arithmetic reasoning. A foundational math benchmark.
HuggingFace evaluation of MATH Level 5 problems. Competition math requiring advanced reasoning and proof construction.
Sentence completion requiring commonsense reasoning about physical and social situations. Tests real-world understanding.
Trivia questions sourced from trivia enthusiasts and quiz websites. Tests breadth of general knowledge.
Language modeling benchmark testing ability to predict the last word of passages requiring long-range context understanding.
- Typetext
- ContextN/A
- ReleasedJan 2024
- LicenseOpen Source
- Statusbenchmark-only