Technology & Science

Microsoft Rolls Out Maia 200: Second-Gen Inference Chip Hits Iowa Datacenter

On 26 Jan 2026 Microsoft began live deployment of its self-designed Maia 200 AI accelerator in Azure’s US-Central region and opened the SDK to developers, marking its first production-scale alternative to Nvidia GPUs for large-language-model inference.

By Priya Castellano

Focusing Facts

  1. Built on TSMC 3 nm with ≈140 billion transistors, Maia 200 delivers >10 PFLOPS FP4 and >5 PFLOPS FP8 within a 750 W envelope.
  2. Each chip integrates a 2.8 TB/s bidirectional Ethernet NIC and can be meshed into clusters of up to 6,144 accelerators; first racks are online in Iowa, with Phoenix slated next.
  3. Microsoft claims 30 % better performance-per-dollar than Maia 100 and 3× the FP4 throughput of AWS Trainium 3.

You've read the facts. The perspectives are behind this line.

Sign up for daily briefings and 5 full articles per week. No credit card.

Perspectives in this article

  • Microsoft corporate communications and company-aligned publications
  • Technology trade and business news media
  • Investor-focused financial commentary
Share

Related Stories