Beta

Google Cloud TPU v6e (Trillium) Pod

Pod / clusterShippingTrillium2025

Google's latest custom AI accelerator in pod configuration. 256 TPU v6e chips connected via custom ICI (Inter-Chip Interconnect). Optimized for both training and inference of Google's Gemini models. Available on Google Cloud as Cloud TPU v6e. Significant perf/watt improvement over v5e.

256
GPUs per system
230 FP8 PFLOPS
GPU model
GT
Google TPU v6e
GPU count
256x
CPU model
Custom host
64x
Memory type
H
HBM3e
Total HBM
8 TB
Host memory
16 TB
Interconnect
ICI 4.0
51.2 TB/s
Networking
1600 Gbps
Storage
100 TB Persistent SSD
Form factor
Multi-rack pod
Weight
TBD
Rack units
TBD

Manufacturer datasheet values · aggregate system compute

FP4 PFLOPSTBD
FP8 PFLOPS230
FP16 PFLOPS115
BF16 PFLOPS115
Training effective PFLOPS90

Thermal envelope · cooling requirements · efficiency

Rack power
60 kW
Per GPU
200 W
Cooling
liquid
PUE estimate
1.1
Power draw relative to tracked systems60 kW / 2500 kW max
3.83 FP8 PFLOPS per kW · average across all systems is 4.81

Hardware amortized over 3 years · power at $0.05/kWh

List price
TBD
Lease monthly
$280,000/mo
Cost per GPU per month
$1,094
TCO per PFLOPS per year
$14,734
PFLOPS per kW
3.83
63% above the average TCO of $9,046/PFLOPS/year across all tracked systems
Available from
GC
Google Cloud

Disclosed in press releases, SEC filings, and conference talks

Quantity
Gemini 2.5 training
Source
Google I/O 2025
Quantity
Public cloud availability
Source
Google Cloud blog

Every data point on this page is reproducible

Compare across the system landscape