AWS detailed its next-generation AI accelerators, confirming Trainium3 Ultra Servers are generally available and previewing Trainium4 for future large-scale training.
Trainium3 uses 3 nm technology, packs 144 chips per rack, delivers hundreds of FP8 petaflops and more than 700 TB/s bandwidth, and achieves multiple-fold improvements in compute, memory bandwidth, and tokens per megawatt over earlier generations.
Over one million Trainium chips are already deployed, making it a multi-billion-dollar business. Trainium4 is designed to further increase FP4 compute and memory bandwidth for the very largest models, reinforcing AWS’s commitment to cost-efficient, high-scale AI infrastructure.





