Better than 9% of all models
Context
N/A
Input $/1M
TBD
Output $/1M
TBD
Type
text
License
Proprietary
Benchmarks
4 tested
Data updated today
About
Tested on 4 benchmarks with 16.6% average. Top scores: GPQA diamond (31.2%), OTIS Mock AIME 2024-2025 (29.9%), ARC-AGI (5.0%).
Capabilities
reasoning
2.5
#175 globally
math
29.9
#133 globally
knowledge
31.2
#171 globally
Benchmark Scores
Compare AllTested on 4 benchmarks · Ranked across 3 categories
Score Distribution (all 231 models)
0255075100
▲ You are here
reasoningCompare reasoning →
ARC-AGI
5.0—Abstraction and Reasoning Corpus. Tests fluid intelligence through novel visual pattern recognition puzzles. Core measure of general intelligence.
ARC-AGI-2
0.1—ARC-AGI 2, harder sequel to ARC. More complex abstract reasoning patterns that test generalization ability beyond training data.
mathCompare math →
OTIS Mock AIME 2024-2025
29.9—Mock AIME (American Invitational Mathematics Exam) problems from OTIS. Tests mathematical competition performance.
knowledgeCompare knowledge →
GPQA diamond
31.2—Graduate-level science questions written by PhD experts. Diamond subset contains questions where experts disagree, testing deep understanding.
Excellent (85+) Good (70-85) Average (50-70) Below (<50)
Links
Info
Research
Documentation
Community
BenchGecko API
magistral-small-1-1
Specifications
- Typetext
- ContextN/A
- ReleasedJan 2024
- LicenseProprietary
- Statusbenchmark-only
Available On
U
UnknownTBDLearn More
Share & Export
Frequently Asked Questions
Magistral Small 1.1 is a proprietary text AI model by Unknown, released in January 2024. It has an average benchmark score of 17.0.