Home/Models/R1 Distill Llama 70B
DeepSeek logo

R1 Distill Llama 70B

by DeepSeek · Released Jan 2025

Open Source
47.6
avg score
Rank #117
Compare
Better than 50% of all models
Context
131K tokens (~66 books)
Input $/1M
$0.70
Output $/1M
$0.80
Type
text
License
Open Source
Benchmarks
8 tested
Data updated today
About

DeepSeek R1 Distill Llama 70B is a distilled large language model based on [Llama-3.3-70B-Instruct](/meta-llama/llama-3.3-70b-instruct), using outputs from [DeepSeek R1](/deepseek/deepseek-r1). The model combines advanced distillation techniques to achieve high performance across...

Tested on 8 benchmarks with 27.8% average. Top scores: IFEval (43.4%), MMLU-PRO (41.6%), BBH (HuggingFace) (35.8%).

Capabilities
reasoning
13.3
#119 globally
math
30.7
#130 globally
knowledge
21.8
#195 globally
speed
22.8
#53 globally
general
35.8
#24 globally
language
43.4
#110 globally
Benchmark Scores
Compare All
Tested on 8 benchmarks · Ranked across 6 categories
Score Distribution (all 233 models)
0255075100
▲ You are here
MUSR

HuggingFace MuSR (Multi-Step Reasoning). Tests multi-hop reasoning requiring chaining multiple facts together.

13.3
MATH Level 5

HuggingFace evaluation of MATH Level 5 problems. Competition math requiring advanced reasoning and proof construction.

30.7
MMLU-PRO

HuggingFace MMLU-Pro. Harder version of MMLU with 10 answer choices instead of 4 and more challenging questions.

41.6
GPQA

HuggingFace evaluation of GPQA (Graduate-Level Google-Proof Q&A). PhD-level science questions that cannot be easily searched.

2.0
Excellent (85+) Good (70-85) Average (50-70) Below (<50)
Links
Documentation
Community
BenchGecko API
deepseek-r1-distill-llama-70b
Specifications
  • Typetext
  • Context131K tokens (~66 books)
  • ReleasedJan 2025
  • LicenseOpen Source
  • StatusActive
  • Cost / Message~$0.002
Available On
DeepSeek logoDeepSeek$0.70
Share & Export
Tweet
R1 Distill Llama 70B is an open-source text AI model by DeepSeek, released in January 2025. It has an average benchmark score of 47.6. Context window: 131K tokens.