Compare · ModelsLive · 2 picked · head to head
R1 Distill Qwen 32B vs Llama 3 8B Instruct
Side by side · benchmarks, pricing, and signals you can act on.
Winner summary
R1 Distill Qwen 32B wins on 4/6 benchmarks
R1 Distill Qwen 32B wins 4 of 6 shared benchmarks. Leads in knowledge · language · math.
Category leads
general·Llama 3 8B Instructknowledge·R1 Distill Qwen 32Blanguage·R1 Distill Qwen 32Bmath·R1 Distill Qwen 32Breasoning·Llama 3 8B Instruct
Hype vs Reality
Attention vs performance
R1 Distill Qwen 32B
#210 by perf·no signal
Llama 3 8B Instruct
#184 by perf·no signal
Best value
Llama 3 8B Instruct
11.1x better value than R1 Distill Qwen 32B
R1 Distill Qwen 32B
79.3 pts/$
$0.29/M
Llama 3 8B Instruct
880.0 pts/$
$0.04/M
Vendor risk
Mixed exposure
One or more vendors flagged
DeepSeek
$3.4B·Tier 1
Meta AI
$1.50T·Tier 1
Head to head
6 benchmarks · 2 models
R1 Distill Qwen 32BLlama 3 8B Instruct
BBH (HuggingFace)
Llama 3 8B Instruct leads by +1.2
R1 Distill Qwen 32B
17.1
Llama 3 8B Instruct
18.4
GPQA
R1 Distill Qwen 32B leads by +2.5
R1 Distill Qwen 32B
4.6
Llama 3 8B Instruct
2.1
IFEval
R1 Distill Qwen 32B leads by +17.8
R1 Distill Qwen 32B
41.9
Llama 3 8B Instruct
24.0
MATH Level 5
R1 Distill Qwen 32B leads by +13.2
R1 Distill Qwen 32B
17.1
Llama 3 8B Instruct
3.9
MMLU-PRO
R1 Distill Qwen 32B leads by +23.2
R1 Distill Qwen 32B
41.0
Llama 3 8B Instruct
17.8
MUSR
Llama 3 8B Instruct leads by +3.8
R1 Distill Qwen 32B
16.1
Llama 3 8B Instruct
19.9
Full benchmark table
| Benchmark | R1 Distill Qwen 32B | Llama 3 8B Instruct |
|---|---|---|
BBH (HuggingFace) | 17.1 | 18.4 |
GPQA | 4.6 | 2.1 |
IFEval | 41.9 | 24.0 |
MATH Level 5 | 17.1 | 3.9 |
MMLU-PRO | 41.0 | 17.8 |
MUSR | 16.1 | 19.9 |
Pricing · per 1M tokens · projected $/mo at 10M tokens
| Model | Input | Output | Context | Projected $/mo |
|---|---|---|---|---|
| $0.29 | $0.29 | 33K tokens (~16 books) | $2.90 | |
| $0.03 | $0.04 | 8K tokens (~4 books) | $0.33 |