Technology & Science
Microsoft Rolls Out Maia 200: Second-Gen Inference Chip Hits Iowa Datacenter
On 26 Jan 2026 Microsoft began live deployment of its self-designed Maia 200 AI accelerator in Azure’s US-Central region and opened the SDK to developers, marking its first production-scale alternative to Nvidia GPUs for large-language-model inference.
Focusing Facts
- Built on TSMC 3 nm with ≈140 billion transistors, Maia 200 delivers >10 PFLOPS FP4 and >5 PFLOPS FP8 within a 750 W envelope.
- Each chip integrates a 2.8 TB/s bidirectional Ethernet NIC and can be meshed into clusters of up to 6,144 accelerators; first racks are online in Iowa, with Phoenix slated next.
- Microsoft claims 30 % better performance-per-dollar than Maia 100 and 3× the FP4 throughput of AWS Trainium 3.
You've read the facts. The perspectives are behind this line.
Perspectives in this article
- Microsoft corporate communications and company-aligned publications
- Technology trade and business news media
- Investor-focused financial commentary