Microsoft Releases Maia 200 Chip to Scale AI Inference and Reduce Power Costs

Microsoft has launched the Maia 200, a custom-designed silicon platform built to scale AI inference with higher performance and improved energy efficiency. The new chip succeeds the Maia 100 introduced in 2023 and features more than 100 billion transistors, delivering over 10 petaflops of 4-bit performance and roughly 5 petaflops at 8-bit precision.

As inference costs become a growing share of AI operating expenses, Microsoft is positioning Maia 200 as a workhorse capable of running today’s largest models while supporting future growth. The chip is already powering workloads from Microsoft’s Superintelligence team and Copilot, and the company has opened access to its software development kit for developers, researchers, and frontier AI labs. The launch underscores Microsoft’s broader strategy to optimize AI infrastructure while reducing reliance on third-party GPUs.

Featured image: Credit: Microsoft

James Dargan

James Dargan is a writer and researcher at The AI Insider. His focus is on the AI startup ecosystem and he writes articles on the space that have a tone accessible to the average reader.

Share this article:

AI Insider

Discover the future of AI technology with "AI Insider" - your go-to platform for industry data, market insights, and groundbreaking AI news

Subscribe today for the latest news about the AI landscape