MiniMax-M2 is a compact, high-efficiency large language model optimized for end-to-end coding and agentic workflows. With 10 billion activated parameters (230 billion total), it delivers near-frontier intelligence across general reasoning,...
Tested on 8 benchmarks with 69.5% average. Top scores: Chatbot Arena Elo — Overall (1346.6%), Chatbot Arena Elo — Coding (1303.3%), OpenCompass — IFEval (90.2%).
OpenCompass Live Code Bench v6. Fresh competitive programming problems to evaluate code generation without memorization.
OpenCompass evaluation on AIME 2025 problems. Tests mathematical reasoning on fresh competition problems.
OpenCompass MMLU-Pro evaluation. Harder knowledge test with more answer choices.
OpenCompass evaluation of GPQA Diamond. PhD-level science questions from the hardest subset.
OpenCompass evaluation of Humanitys Last Exam. Expert-level cross-discipline knowledge test.
- Typetext
- Context197K tokens (~98 books)
- ReleasedOct 2025
- LicenseOpen Source
- StatusActive
- Cost / Message~$0.002