Beta
Compute · ChipsAnchor of the compute layer

Every AI Chip · Tracked

Every AI chip from every manufacturer and every foundry. Specs, power, cloud pricing, known buyers, and live supply tightness · cross-linked to foundries, memory suppliers, racks, datacenters, and the models trained on them.

Supply trackedFab-level originHBM supplier mapPerf per Watt
Preview · full AICDI on /compute
AI Compute Demand Index
+12 WoW
FULL PARABOLIC
No bubble in sight · full parabolic. 100 = balanced supply · unbounded above · a reading of 250+ means no bubble in sight.
How it's built
  • Weighted tightness68
  • Lead time multiplier2.09×
  • Capex accelerator1.85×
  • Frontier holdout premium+60
0100 balanced300+
20
Tracked chips
10
Frontier tier
10
Manufacturers
90%
TSMC share
20
Shipping now
8
On HBM3e
25
Disclosed buyer entries
23
Known models trained

Frontier · Mainstream · Specialized · Legacy · assessed quarterly

Frontier
10 · 50%

Current-generation flagship silicon shipping at scale to hyperscalers. The chips training the biggest models right now.

Mainstream
5 · 25%

Current-gen non-flagship or last-gen flagship still deployed widely. The workhorses of the AI economy.

Specialized
4 · 20%

Wafer-scale, custom hyperscaler silicon, or experimental architectures serving niche workloads.

Legacy
1 · 5%

Older chips being phased out or already past EOL. Still used in research clusters and price-sensitive workloads.

FP16 TFLOPs vs TDP · log scale · bubble size = HBM capacity

0.5 T/W1 T/W2 T/W4 T/W
0.5 T/W1 T/W2 T/W4 T/W1003007001.5k3k10kTDP (Watts) · log1005002k10k50k200kFP16 TFLOPs · logTPU v7WSE-3TPU v6eB200B300GB200MI355XGB300Ascend 910CTrainium2MI325X

Up-and-to-the-left wins. Chips above the 4 T/W line are the efficiency frontier — that's where the power-budget game is won at hyperscale. Wafer-scale and multi-die superchips sit top-right because they trade watts for raw throughput.

How tight is supply right now · hand-curated from order books + earnings calls

sold out tight glut
NVIDIA logoGB200
Sold out98
52wk lead
NVIDIA logoGB300
Sold out95
48wk lead
NVIDIA logoB200
Sold out92
38wk lead
Huawei logoAscend 910C
Sold out92
44wk lead
Cerebras logoWSE-3
Sold out90
internal
NVIDIA logoB300
Tight88
36wk lead
AMD logoMI355X
Tight80
30wk lead
AWS logoTrainium2
Tight75
24wk lead
AMD logoMI325X
Tight70
22wk lead
Microsoft logoMaia 100
Tight70
internal
NVIDIA logoH200
Balanced65
16wk lead
Google logoTPU v7
Balanced60
internal
Groq logoLPU
Balanced60
internal
Google logoTPU v6e
Balanced55
internal
Meta logoMTIA v2
Balanced50
internal
AMD logoMI300X
Balanced40
6wk lead
NVIDIA logoH100
Loose35
4wk lead
Intel logoGaudi 3
Loose25
4wk lead
Google logoTPU v5p
Loose25
internal
NVIDIA logoA100
Glut15
2wk lead

Supply notes

  • GB200Sold out through 2027 · Oracle, Microsoft, Meta competing
  • GB300xAI Memphis + Stargate phase 2
  • B200CoWoS-L packaging capacity is the bottleneck
  • Ascend 910CChina export-ban alternative · SMIC N7 constrained
  • WSE-3G42 Condor Galaxy absorbing most wafers
  • B300Ultra variant ramping Q4 2025

This is the per-chip preview of the AI Compute Demand Index landing on /compute · unbounded above 100, goes to 250+ when the market is on fire. Updated weekly from hyperscaler order books and earnings disclosures.

Six curated shortlists computed from the live spec data

Top 5 by FP16/BF16 TFLOPs · ignore price and power · pure compute

  1. 1Cerebras logoWSE-3Cerebras
  2. 2NVIDIA logoGB300NVIDIA
  3. 3Google logoTPU v7Google
  4. 4NVIDIA logoGB200NVIDIA
  5. 5NVIDIA logoB300NVIDIA

Best FP16 TFLOPs per Watt · the chips that make datacenter power budgets work

  1. 1Google logoTPU v7Google
  2. 2Cerebras logoWSE-3Cerebras
  3. 3Meta logoMTIA v2Meta
  4. 4Google logoTPU v6eGoogle
  5. 5NVIDIA logoB200NVIDIA

Biggest memory for fitting 400B+ parameter models in a single node

  1. 1NVIDIA logoGB300NVIDIA
  2. 2NVIDIA logoGB200NVIDIA
  3. 3AMD logoMI355XAMD
  4. 4NVIDIA logoB300NVIDIA
  5. 5AMD logoMI325XAMD

Lowest $/h in public clouds · best for quick experiments

  1. 1NVIDIA logoA100NVIDIA
  2. 2AWS logoTrainium2AWS
  3. 3NVIDIA logoH100NVIDIA
  4. 4Google logoTPU v6eGoogle
  5. 5AMD logoMI300XAMD

The chips training the biggest frontier models right now

  1. 1NVIDIA logoB200NVIDIA
  2. 2NVIDIA logoGB200NVIDIA
  3. 3Google logoTPU v7Google
  4. 4AWS logoTrainium2AWS
  5. 5NVIDIA logoGB300NVIDIA

Wafer-scale, LPU, MTIA and other custom silicon beyond GPU/TPU

  1. 1Cerebras logoWSE-3Cerebras
  2. 2Meta logoMTIA v2Meta
  3. 3Microsoft logoMaia 100Microsoft
  4. 4Groq logoLPUGroq

Sorted by Gecko score · composite of tier, recency, adoption, perf density

#ChipMakerScore
1NVIDIA94
2NVIDIA86
3Google83
4AWS78
5NVIDIA75
6AMD75
7Huawei75
8Google71
9NVIDIA70
10AMD68
11AMD65
12Cerebras60
13Intel57
14NVIDIA55
15NVIDIA55
16Meta52
17Google49
18Microsoft42
19Groq34
20NVIDIA31

10 companies · 3 foundries · 2 countries

Cerebras logoCerebrasUSA flag
Chips
1
Frontier
0
Avg T/W
5.43
Top: WSE-3
USA flagUSA
NVIDIA logoNVIDIAUSA flag
Chips
7
Frontier
4
Avg T/W
1.55
Top: B200
USA flagUSA
Google logoGoogleUSA flag
Chips
3
Frontier
2
Avg T/W
3.78
Top: TPU v7
USA flagUSA
AMD logoAMDUSA flag
Chips
3
Frontier
2
Avg T/W
1.56
Top: MI355X
USA flagUSA
Intel logoIntelUSA flag
Chips
1
Frontier
0
Avg T/W
2.04
Top: Gaudi 3
USA flagUSA
Huawei logoHuaweiChina flag
Chips
1
Frontier
1
Avg T/W
1.45
Top: Ascend 910C
China flagChina
Microsoft logoMicrosoftUSA flag
Chips
1
Frontier
0
Avg T/W
1.14
Top: Maia 100
USA flagUSA
AWS logoAWSUSA flag
Chips
1
Frontier
1
Avg T/W
1.33
Top: Trainium2
USA flagUSA
Meta logoMetaUSA flag
Chips
1
Frontier
0
Avg T/W
3.93
Top: MTIA v2
USA flagUSA
Groq logoGroqUSA flag
Chips
1
Frontier
0
Avg T/W
0.5
Top: LPU
USA flagUSA

Pulled from the live dataset · schema-ready for AEO

Which AI chip has the most FP16 TFLOPs right now?

NVIDIA GB300 (Grace Blackwell Ultra) leads dense FP16 throughput at ~5000 TFLOPs per superchip, followed by GB200 at 4500 TFLOPs. The Cerebras WSE-3 reaches 125000 TFLOPs per wafer but is a fundamentally different architecture · a single wafer contains what would otherwise be ~50 GPUs.

Which AI chip is most efficient per Watt?

Google TPU v7 (Ironwood) and NVIDIA B200 lead FP16 TFLOPs per Watt among shipping chips. Trillium (TPU v6e) is close behind. Inference-first designs (Ironwood, Groq LPU) tend to dominate perf/Watt because they skip training-specific circuitry.

Who manufactures AI chips besides NVIDIA?

10 companies ship production AI accelerators tracked on BenchGecko · NVIDIA, AMD, Google, AWS, Intel, Microsoft, Meta, Huawei, Cerebras, Groq, and more. NVIDIA dominates third-party sales but hyperscalers are shipping more custom silicon (Trainium2, TPU v7, Maia 100, MTIA) to reduce dependence.

Which fabs make AI chips?

90% of tracked chips are fabbed at TSMC. GlobalFoundries handles Groq's 14nm LPU. SMIC fabricates Huawei's Ascend 910C at their N7 node as the primary answer to the U.S. export ban. TSMC advanced nodes (N5/N4/N3) are the bottleneck for current-gen NVIDIA, AMD, AWS, Google, and Microsoft silicon.

How do you decide which chip is "frontier" tier?

Frontier means current-generation flagship, shipping in 2025-2026, widely ordered by hyperscalers. Mainstream means current-gen non-flagship or last-gen flagship still deployed (H100, H200, MI300X). Specialized means wafer-scale, LPU, MTIA and other custom designs. Legacy means EOL or phasing out (A100, V100). Tiers are reviewed quarterly.

Where does BenchGecko get chip data from?

Manufacturer datasheets for specs, SEC 10-K filings and earnings transcripts for buyer disclosures, cloud provider API pricing for $/hour, public press releases for release dates. Every chip detail page includes source links. No paywalled data · no vendor-sponsored rankings · everything is reproducible.

Keep exploring the compute graph