Compare · ModelsLive · 2 picked · head to head
Qwen2 7B Instruct vs Meta Llama 3 8B Instruct
Side by side · benchmarks, pricing, and signals you can act on.
Winner summary
Qwen2 7B Instruct wins on 22/25 benchmarks
Qwen2 7B Instruct wins 22 of 25 shared benchmarks. Leads in general · knowledge · language.
Category leads
general·Qwen2 7B Instructknowledge·Qwen2 7B Instructlanguage·Qwen2 7B Instructmath·Qwen2 7B Instructreasoning·Qwen2 7B Instruct
Hype vs Reality
Attention vs performance
Qwen2 7B Instruct
#92 by perf·no signal
Meta Llama 3 8B Instruct
#113 by perf·no signal
Best value
Pricing unknown
Qwen2 7B Instruct
—
no price
Meta Llama 3 8B Instruct
—
no price
Vendor risk
Who is behind the model
Alibaba (Qwen)
$293.0B·Tier 1
Meta AI
$1.50T·Tier 1
Head to head
25 benchmarks · 2 models
Qwen2 7B InstructMeta Llama 3 8B Instruct
BBH (HuggingFace)
Qwen2 7B Instruct leads by +9.6
Qwen2 7B Instruct
37.8
Meta Llama 3 8B Instruct
28.2
GPQA
Qwen2 7B Instruct leads by +5.2
Qwen2 7B Instruct
6.4
Meta Llama 3 8B Instruct
1.2
IFEval
Meta Llama 3 8B Instruct leads by +17.3
Qwen2 7B Instruct
56.8
Meta Llama 3 8B Instruct
74.1
MATH Level 5
Qwen2 7B Instruct leads by +19.0
Qwen2 7B Instruct
27.6
Meta Llama 3 8B Instruct
8.7
MMLU-PRO
Qwen2 7B Instruct leads by +2.0
Qwen2 7B Instruct
31.6
Meta Llama 3 8B Instruct
29.6
MUSR
Qwen2 7B Instruct leads by +5.8
Qwen2 7B Instruct
7.4
Meta Llama 3 8B Instruct
1.6
JCommonsenseQA
Qwen2 7B Instruct leads by +1.4
Qwen2 7B Instruct
89.1
Meta Llama 3 8B Instruct
87.7
JMMLU
Qwen2 7B Instruct leads by +9.8
Qwen2 7B Instruct
56.5
Meta Llama 3 8B Instruct
46.7
JNLI
Qwen2 7B Instruct leads by +20.2
Qwen2 7B Instruct
81.3
Meta Llama 3 8B Instruct
61.1
JSQuAD
Qwen2 7B Instruct leads by +0.1
Qwen2 7B Instruct
89.6
Meta Llama 3 8B Instruct
89.5
LLM-JP · Overall
Qwen2 7B Instruct leads by +2.1
Qwen2 7B Instruct
51.7
Meta Llama 3 8B Instruct
49.6
MMMLU · Arabic
Qwen2 7B Instruct leads by +10.2
Qwen2 7B Instruct
50.7
Meta Llama 3 8B Instruct
40.5
MMMLU · Bengali
Qwen2 7B Instruct leads by +7.0
Qwen2 7B Instruct
43.4
Meta Llama 3 8B Instruct
36.4
MMMLU · German
Qwen2 7B Instruct leads by +3.6
Qwen2 7B Instruct
57.1
Meta Llama 3 8B Instruct
53.5
MMMLU · Spanish
Qwen2 7B Instruct leads by +4.4
Qwen2 7B Instruct
60.2
Meta Llama 3 8B Instruct
55.8
MMMLU · French
Qwen2 7B Instruct leads by +5.0
Qwen2 7B Instruct
60.8
Meta Llama 3 8B Instruct
55.8
MMMLU · Hindi
Qwen2 7B Instruct leads by +3.7
Qwen2 7B Instruct
45.1
Meta Llama 3 8B Instruct
41.4
MMMLU · Indonesian
Qwen2 7B Instruct leads by +3.1
Qwen2 7B Instruct
54.1
Meta Llama 3 8B Instruct
51.0
MMMLU · Italian
Qwen2 7B Instruct leads by +5.7
Qwen2 7B Instruct
59.0
Meta Llama 3 8B Instruct
53.3
MMMLU · Japanese
Qwen2 7B Instruct leads by +14.3
Qwen2 7B Instruct
56.6
Meta Llama 3 8B Instruct
42.3
MMMLU · Korean
Qwen2 7B Instruct leads by +7.5
Qwen2 7B Instruct
54.0
Meta Llama 3 8B Instruct
46.5
MMMLU · Portuguese
Qwen2 7B Instruct leads by +4.6
Qwen2 7B Instruct
60.1
Meta Llama 3 8B Instruct
55.5
MMMLU · Swahili
Meta Llama 3 8B Instruct leads by +3.2
Qwen2 7B Instruct
34.3
Meta Llama 3 8B Instruct
37.5
MMMLU · Yoruba
Meta Llama 3 8B Instruct leads by +0.8
Qwen2 7B Instruct
30.2
Meta Llama 3 8B Instruct
31.0
MMMLU · Chinese
Qwen2 7B Instruct leads by +10.4
Qwen2 7B Instruct
61.8
Meta Llama 3 8B Instruct
51.4
Full benchmark table
| Benchmark | Qwen2 7B Instruct | Meta Llama 3 8B Instruct |
|---|---|---|
BBH (HuggingFace) | 37.8 | 28.2 |
GPQA | 6.4 | 1.2 |
IFEval | 56.8 | 74.1 |
MATH Level 5 | 27.6 | 8.7 |
MMLU-PRO | 31.6 | 29.6 |
MUSR | 7.4 | 1.6 |
JCommonsenseQA | 89.1 | 87.7 |
JMMLU | 56.5 | 46.7 |
JNLI | 81.3 | 61.1 |
JSQuAD | 89.6 | 89.5 |
LLM-JP · Overall | 51.7 | 49.6 |
MMMLU · Arabic | 50.7 | 40.5 |
MMMLU · Bengali | 43.4 | 36.4 |
MMMLU · German | 57.1 | 53.5 |
MMMLU · Spanish | 60.2 | 55.8 |
MMMLU · French | 60.8 | 55.8 |
MMMLU · Hindi | 45.1 | 41.4 |
MMMLU · Indonesian | 54.1 | 51.0 |
MMMLU · Italian | 59.0 | 53.3 |
MMMLU · Japanese | 56.6 | 42.3 |
MMMLU · Korean | 54.0 | 46.5 |
MMMLU · Portuguese | 60.1 | 55.5 |
MMMLU · Swahili | 34.3 | 37.5 |
MMMLU · Yoruba | 30.2 | 31.0 |
MMMLU · Chinese | 61.8 | 51.4 |
Pricing · per 1M tokens · projected $/mo at 10M tokens
| Model | Input | Output | Context | Projected $/mo |
|---|---|---|---|---|
| — | — | — | — | |
| — | — | — | — |