Compare · ModelsLive · 3 picked · head to head
Qwen2 VL 7B Instruct vs DeepSeek R1 Distill Qwen 14B vs Qwen2 7B Instruct
Side by side · benchmarks, pricing, and signals you can act on.
Winner summary
DeepSeek R1 Distill Qwen 14B wins on 9/11 benchmarks
DeepSeek R1 Distill Qwen 14B wins 9 of 11 shared benchmarks. Leads in general · knowledge · language.
Category leads
general·DeepSeek R1 Distill Qwen 14Bknowledge·DeepSeek R1 Distill Qwen 14Blanguage·DeepSeek R1 Distill Qwen 14Bmath·DeepSeek R1 Distill Qwen 14Breasoning·DeepSeek R1 Distill Qwen 14B
Hype vs Reality
Attention vs performance
Qwen2 VL 7B Instruct
#106 by perf·no signal
DeepSeek R1 Distill Qwen 14B
#64 by perf·no signal
Qwen2 7B Instruct
#94 by perf·no signal
Best value
Pricing unknown
Qwen2 VL 7B Instruct
—
no price
DeepSeek R1 Distill Qwen 14B
—
no price
Qwen2 7B Instruct
—
no price
Vendor risk
Mixed exposure
One or more vendors flagged
Alibaba (Qwen)
$293.0B·Tier 1
DeepSeek
$3.4B·Tier 1
Alibaba (Qwen)
$293.0B·Tier 1
Head to head
11 benchmarks · 3 models
Qwen2 VL 7B InstructDeepSeek R1 Distill Qwen 14BQwen2 7B Instruct
BBH (HuggingFace)
DeepSeek R1 Distill Qwen 14B leads by +2.9
Qwen2 VL 7B Instruct
35.9
DeepSeek R1 Distill Qwen 14B
40.7
Qwen2 7B Instruct
37.8
GPQA
DeepSeek R1 Distill Qwen 14B leads by +9.1
Qwen2 VL 7B Instruct
9.3
DeepSeek R1 Distill Qwen 14B
18.3
Qwen2 7B Instruct
6.4
IFEval
Qwen2 7B Instruct leads by +10.8
Qwen2 VL 7B Instruct
46.0
DeepSeek R1 Distill Qwen 14B
43.8
Qwen2 7B Instruct
56.8
MATH Level 5
DeepSeek R1 Distill Qwen 14B leads by +29.4
Qwen2 VL 7B Instruct
19.9
DeepSeek R1 Distill Qwen 14B
57.0
Qwen2 7B Instruct
27.6
MMLU-PRO
DeepSeek R1 Distill Qwen 14B leads by +6.4
Qwen2 VL 7B Instruct
34.4
DeepSeek R1 Distill Qwen 14B
40.7
Qwen2 7B Instruct
31.6
MUSR
DeepSeek R1 Distill Qwen 14B leads by +15.2
Qwen2 VL 7B Instruct
13.6
DeepSeek R1 Distill Qwen 14B
28.7
Qwen2 7B Instruct
7.4
JCommonsenseQA
DeepSeek R1 Distill Qwen 14B leads by +4.6
Qwen2 VL 7B Instruct
87.8
DeepSeek R1 Distill Qwen 14B
93.7
Qwen2 7B Instruct
89.1
JMMLU
DeepSeek R1 Distill Qwen 14B leads by +6.9
Qwen2 VL 7B Instruct
56.3
DeepSeek R1 Distill Qwen 14B
63.4
Qwen2 7B Instruct
56.5
JNLI
DeepSeek R1 Distill Qwen 14B leads by +1.1
Qwen2 VL 7B Instruct
74.4
DeepSeek R1 Distill Qwen 14B
82.4
Qwen2 7B Instruct
81.3
JSQuAD
Qwen2 VL 7B Instruct leads by +0.1
Qwen2 VL 7B Instruct
89.9
DeepSeek R1 Distill Qwen 14B
89.8
Qwen2 7B Instruct
89.6
LLM-JP · Overall
DeepSeek R1 Distill Qwen 14B leads by +3.9
Qwen2 VL 7B Instruct
53.0
DeepSeek R1 Distill Qwen 14B
56.8
Qwen2 7B Instruct
51.7
Full benchmark table
| Benchmark | Qwen2 VL 7B Instruct | DeepSeek R1 Distill Qwen 14B | Qwen2 7B Instruct |
|---|---|---|---|
BBH (HuggingFace) | 35.9 | 40.7 | 37.8 |
GPQA | 9.3 | 18.3 | 6.4 |
IFEval | 46.0 | 43.8 | 56.8 |
MATH Level 5 | 19.9 | 57.0 | 27.6 |
MMLU-PRO | 34.4 | 40.7 | 31.6 |
MUSR | 13.6 | 28.7 | 7.4 |
JCommonsenseQA | 87.8 | 93.7 | 89.1 |
JMMLU | 56.3 | 63.4 | 56.5 |
JNLI | 74.4 | 82.4 | 81.3 |
JSQuAD | 89.9 | 89.8 | 89.6 |
LLM-JP · Overall | 53.0 | 56.8 | 51.7 |
Pricing · per 1M tokens · projected $/mo at 10M tokens
| Model | Input | Output | Context | Projected $/mo |
|---|---|---|---|---|
| — | — | — | — | |
| — | — | — | — | |
| — | — | — | — |