Microsoft Maia 100 AI Accelerator Rack
Full rackRampingMaia 1002025
Microsoft's first custom AI silicon at rack scale. Maia 100 chips fabricated at TSMC on N5 with HBM3e. Designed for Azure AI inference and fine-tuning. Co-designed with the Cobalt ARM CPU for host compute. Microsoft's strategic hedge against NVIDIA dependency.
32
GPUs per system
96 FP8 PFLOPS
Total HBM
6.1 TB
Host memory
8 TB
Interconnect
Custom Ethernet fabric
8 TB/s
Networking
1600 Gbps
Storage
60 TB NVMe SSD
Form factor
Custom rack
Weight
TBD
Rack units
42U
Performance
Manufacturer datasheet values · aggregate system compute
| FP4 PFLOPS | TBD |
| FP8 PFLOPS | 96 |
| FP16 PFLOPS | 48 |
| BF16 PFLOPS | 48 |
| Training effective PFLOPS | 36 |
Power and cooling
Thermal envelope · cooling requirements · efficiency
Rack power
40 kW
Per GPU
500 W
Cooling
liquid
PUE estimate
1.1
Power draw relative to tracked systems40 kW / 2500 kW max
2.4 FP8 PFLOPS per kW · average across all systems is 4.81
TCO analysis
Hardware amortized over 3 years · power at $0.05/kWh
List price
TBD
PFLOPS per kW
2.4
Available from
MA
Microsoft AzureKnown deployments
Disclosed in press releases, SEC filings, and conference talks
Quantity
Internal deployment
Source
Microsoft Ignite 2024Sources
Every data point on this page is reproducible
Other AI systems
Compare across the system landscape
8960x Google TPU v5p · 8100 PFLOPS
Pod / clusterShipping
72x NVIDIA B300 · 1440 PFLOPS
Full rackAnnounced
72x NVIDIA B200 · 720 PFLOPS
Full rackShipping
256x Google TPU v6e · 230 PFLOPS
Pod / clusterShipping
8x NVIDIA B200 · 80 PFLOPS
Server nodeShipping
8x AMD MI325X · 48 PFLOPS
Server nodeShipping