By using this site, you agree to our Privacy Policy and Terms of Use.
Accept
VellaTimesVellaTimesVellaTimes
  • News
    NewsShow More
    A glowing antimatter atom passing through a hexagonal graphene sheet and splitting into a quantum wave interference pattern in a high-tech laboratory setting.
    Scientists Observe Positronium Wave Behavior in Lab
    May 1, 2026
    Hyper-realistic news-style image of a modern AI data center with server racks and a digital display labeled DeepSeek V4, shown in cool blue lighting.
    DeepSeek V4 launch puts Huawei AI chips in spotlight
    May 1, 2026
    A modern smartphone displaying an app storefront positioned next to a wooden judge's gavel on a desk, representing the legal battle over digital marketplace policies.
    Apple Loses Bid to Pause App Store Fee Changes
    May 1, 2026
    The NASA Curiosity rover is using its robotic arm to drill into a red sandstone rock on the dusty surface of Mars.
    Mars Organic Molecules: Curiosity Rover Makes Historic Find
    May 1, 2026
    News-style image of Elon Musk seated in a courtroom during a legal dispute involving OpenAI.
    Elon Musk OpenAI Trial Puts Nonprofit Mission on Trial
    May 1, 2026
  • Technology
    TechnologyShow More
    A modern smartphone displaying an app storefront positioned next to a wooden judge's gavel on a desk, representing the legal battle over digital marketplace policies.
    Apple Loses Bid to Pause App Store Fee Changes
    May 1, 2026
    A business professional using an AI assistant on a laptop in a modern office with a data center visible in the background.
    Microsoft Copilot Tops 20 Million Paid Enterprise Seats
    May 1, 2026
    A brightly lit modern semiconductor cleanroom featuring advanced silicon wafers and glowing blue server racks.
    Samsung Q1 Profit Surges Eightfold as AI Boom Fuels Record Chip Earnings
    April 30, 2026
    A person holding a smartphone displaying the Amazon Shopping app's AI audio chat interface in a modern living room.
    Amazon AI Audio Shopping Chat Enhanced With Real-Time Q&A
    April 29, 2026
    Hyper-realistic news-style image of a modern cloud operations room with Amazon Bedrock and OpenAI-themed screens and a coding dashboard in view.
    OpenAI Models and Codex Debut on Amazon Bedrock in Preview
    April 29, 2026
  • AI
    AIShow More
    Hyper-realistic news-style image of a modern AI data center with server racks and a digital display labeled DeepSeek V4, shown in cool blue lighting.
    DeepSeek V4 launch puts Huawei AI chips in spotlight
    May 1, 2026
    News-style image of Elon Musk seated in a courtroom during a legal dispute involving OpenAI.
    Elon Musk OpenAI Trial Puts Nonprofit Mission on Trial
    May 1, 2026
    News-style image showing LG Electronics and Nvidia branding in a modern tech setting with AI server racks and a service robot.
    Nvidia-LG Talks Highlight Wider AI Expansion Strategy
    April 30, 2026
    A dramatic courtroom setting featuring an abstract artificial intelligence hologram on a wooden table, representing the high-stakes tech trial.
    Elon Musk vs Sam Altman OpenAI Trial Over AI Future
    April 29, 2026
    A glowing artificial intelligence processor chip with V4 etched on the surface, mounted on an illuminated circuit board.
    DeepSeek V4 Launches on Huawei Chips for AI Innovation
    April 29, 2026
  • Science
    ScienceShow More
    A glowing antimatter atom passing through a hexagonal graphene sheet and splitting into a quantum wave interference pattern in a high-tech laboratory setting.
    Scientists Observe Positronium Wave Behavior in Lab
    May 1, 2026
    The NASA Curiosity rover is using its robotic arm to drill into a red sandstone rock on the dusty surface of Mars.
    Mars Organic Molecules: Curiosity Rover Makes Historic Find
    May 1, 2026
    Aerial view of the Pacific Ocean off a forested coastline with a glowing geological fault line beneath the water representing the Cascadia subduction zone.
    Earth Tearing Apart Under the Cascadia Subduction Zone
    May 1, 2026
    A young adult female patient and a doctor are looking at medical charts in a modern clinical office setting.
    Rising Cancer Rates in Young Adults: Is Obesity to Blame?
    April 29, 2026
    Microscopic view of engineered immune cells glowing in a modern, high-tech medical laboratory setting.
    CAR-T Cell Therapy Eradicates Severe Autoimmune Diseases
    April 29, 2026
  • World
    WorldShow More
    Allu Arjun Commitment to Ethical Brand Partnerships
    Exploring Allu Arjun’s Commitment to Ethical Brand Partnerships
    December 18, 2023
    Orry aka Orhan Awatramani
    Orhan Awatramani ‘Orry’ Biography, Lifestyle and Rise to Fame
    December 8, 2023
    Alia Bhatt Latest Deepake Video Victim
    Alia Bhatt becomes latest victim of Deepfake Videos, Obscene Video goes Viral
    November 28, 2023
    Napoleon Movie Review
    Napoleon Movie Review: A Historical Epic by Ridley Scott Reviewed
    November 25, 2023
  • Bookmarks
Search
Category
  • News
  • Technology
  • AI
  • Science
  • World
Company
  • About Us
  • Contact Us
  • Fact Checking Policy
  • Terms & Conditions
  • Privacy Policy
  • Copyright Policy
Resources
  • Home
  • Web Stories
  • Bookmarks
  • Interests
  • Disclaimer
  • Sitemap
© 2022 VellaTimes • All Rights Reserved.
Reading: Microsoft Maia 200 AI chip boosts Azure inference
Share
Notification Show More
Font ResizerAa
VellaTimesVellaTimes
Font ResizerAa
  • News
  • Technology
  • AI
  • Science
  • World
Search
  • Explore
    • News
    • Technology
    • AI
    • Science
    • World
  • Useful Links
    • About Us
    • Contact Us
    • Fact Checking Policy
    • Terms & Conditions
    • Privacy Policy
    • Copyright Policy
  • Home
  • Web Stories
  • Bookmarks
  • Interests
  • Disclaimer
  • Sitemap
© 2022 VellaTimes • All Rights Reserved.
News

Microsoft Maia 200 AI chip boosts Azure inference

Rakesh Paul
Last updated: 27/01/2026
Rakesh Paul
Share
8 Min Read
Hyper-realistic image of a modern data center with glowing server racks and a Maia 200 AI accelerator card highlighted in the foreground.

Microsoft has unveiled its Maia 200 AI accelerator, a custom chip built to make running large AI models faster and more cost-efficient across its Azure cloud. The company is positioning Maia 200 as a breakthrough for AI inference, the stage where trained models answer user prompts and power real-world applications like chatbots and productivity tools.

Contents
A custom AI chip focused on inferencePerformance numbers and comparisonsMemory, networking and system designRole in Azure and Microsoft AI servicesDeveloper tools and future roadmap

Maia 200 is designed specifically for large-scale inference rather than training, and Microsoft says it delivers better performance per dollar than the hardware it currently uses in its data centers. It will power services such as Microsoft 365 Copilot, Azure AI Foundry, and models from the Microsoft Superintelligence team, including OpenAI’s latest GPT‑5.2 family.

A custom AI chip focused on inference

Microsoft describes Maia 200 as a “breakthrough inference accelerator” tuned for the heavy workloads of modern reasoning and language models. Unlike some rival chips that are built to handle both training and inference, Microsoft and industry analysts say this design is optimized for the production side of AI, where efficiency and throughput matter most.

The chip is fabricated on Taiwan Semiconductor Manufacturing Company’s 3‑nanometer process and contains more than 140 billion transistors. According to Microsoft, Maia 200 is its most performant first‑party silicon yet and the most efficient inference system it has deployed, delivering 30% better performance per dollar than the latest generation of hardware in its fleet.

Performance numbers and comparisons

Each Maia 200 accelerator can deliver more than 10 petaFLOPS of compute at 4‑bit precision (FP4) and over 5 petaFLOPS at 8‑bit precision (FP8), within a 750‑watt system‑on‑chip power envelope. Microsoft says this is enough for a single chip to run today’s largest AI models while still leaving room for even bigger models in the future.

Microsoft also draws direct comparisons with rival cloud providers’ in‑house AI chips. The company claims Maia 200 delivers three times the FP4 performance of Amazon’s third‑generation Trainium and FP8 performance above Google’s seventh‑generation TPU. Analysts note that Maia 200 uses a more advanced 3‑nanometer manufacturing node than the 5‑nanometer or 7‑nanometer processes used in these competing chips, and say it shows Microsoft is closing earlier gaps in custom silicon.

Microsoft and external commentators emphasize that customers will still need to validate real‑world performance and pricing within Azure before shifting workloads from other vendors, including Nvidia. One analyst also points out that enterprises will want to see how much of Microsoft’s own infrastructure savings from Maia 200 are eventually reflected in cloud subscription costs.

Memory, networking and system design

Maia 200’s architecture centers on feeding data to AI models as efficiently as possible, not just pushing raw compute power. Each chip includes 216GB of high‑bandwidth HBM3e memory delivering 7 TB/s of bandwidth, along with 272MB of on‑chip SRAM and specialized data‑movement engines to keep large models highly utilized.

Microsoft redesigned the memory subsystem around low‑precision data types, a dedicated direct memory access engine, on‑die SRAM, and a custom network‑on‑chip fabric to move data quickly and increase token throughput during inference. At the system level, each accelerator exposes 2.8 TB/s of bidirectional scale‑up bandwidth and connects into a two‑tier network built on standard Ethernet rather than proprietary fabrics.

Clusters can scale up to 6,144 Maia 200 accelerators, using the same communication protocols within trays, racks, and across the data center. Within each tray, four accelerators are directly linked to keep high‑bandwidth communication local, which Microsoft says helps reduce power use and total cost of ownership for dense inference deployments.

Role in Azure and Microsoft AI services

Maia 200 is already deployed in Microsoft’s US Central data center region near Des Moines, Iowa, and is rolling out next to the US West 3 region near Phoenix, Arizona, with more regions planned later. The chip is integrated with Azure as part of a heterogeneous AI infrastructure that also includes other types of accelerators.

Microsoft says Maia 200 will run multiple models, including OpenAI’s GPT‑5.2 family, and will support workloads for Azure AI Foundry and Microsoft 365 Copilot. The Microsoft Superintelligence team plans to use the chip for reinforcement learning and synthetic data generation, which are key for improving future in‑house AI models and speeding up the creation of domain‑specific training data.

Industry analysts argue that Microsoft’s long experience with enterprise IT gives it an advantage in embedding Maia‑based inference services directly into the broader Azure platform. Commentators also stress that Microsoft’s strategy is to complement, not outright replace, other vendors like Nvidia and AMD, while offering customers more options for high‑throughput, memory‑intensive AI inference.

Developer tools and future roadmap

To encourage early adoption, Microsoft is offering a preview of the Maia software development kit. The SDK supports popular AI frameworks, including PyTorch, and includes a Triton compiler, an optimized kernel library, access to Maia’s low‑level NPL programming language, a simulator, and a cost calculator for tuning workloads.

Microsoft says its silicon program used a sophisticated pre‑silicon environment to model large language model workloads and validate networking and cooling systems, including a second‑generation closed‑loop liquid cooling unit. According to the company, this approach allowed AI models to run on Maia 200 within days of first silicon and cut the time from first chip to data center deployment by more than half compared with earlier infrastructure programs.

The company describes Maia as a multi‑generation accelerator family and says it is already designing future versions while it deploys Maia 200 across its global infrastructure. As these chips scale out, Microsoft expects continued improvements in performance per dollar and per watt for its most important AI workloads in Azure.

TAGGED: AI chips, AI inference, Azure, cloud computing, data centers, GPT-5.2, Maia 200, Microsoft
Share This Article
Facebook Twitter Whatsapp Whatsapp Telegram Copy Link
By Rakesh Paul
I'm the Co-Founder of VellaTimes and an experienced digital marketer. With substantial experience in the blogging industry, I love crafting insightful and engaging news articles on technology, sports, and automobiles.
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *


Most Read

Nvidia H200 chips China Buyers asked to pay upfront

January 10, 2026

Entangled atomic clouds boost precision quantum sensing

January 27, 2026

Apple WWDC 2026: June 8 Event, iOS 27, and AI Updates

March 24, 2026

Google Enhances Gemini Workspace Integration With AI Tools

March 11, 2026

TikTok Local Feed Launches in U.S. Using Precise Location

February 12, 2026

Nvidia halts H200 production for China, shifts to Vera Rubin

March 5, 2026

Related News

A glowing antimatter atom passing through a hexagonal graphene sheet and splitting into a quantum wave interference pattern in a high-tech laboratory setting.
News

Scientists Observe Positronium Wave Behavior in Lab

Nisha Pradhan Nisha Pradhan May 1, 2026
Hyper-realistic news-style image of a modern AI data center with server racks and a digital display labeled DeepSeek V4, shown in cool blue lighting.
News

DeepSeek V4 launch puts Huawei AI chips in spotlight

Sameer Katoch Sameer Katoch May 1, 2026
A modern smartphone displaying an app storefront positioned next to a wooden judge's gavel on a desk, representing the legal battle over digital marketplace policies.
News

Apple Loses Bid to Pause App Store Fee Changes

Rakesh Paul Rakesh Paul May 1, 2026

About Us

VellaTimesVellaTimesVellaTimes

VellaTimes is a leading news portal that covers the latest trending news in technology, lifestyle, entertainment, automobiles, travel, and sports.

Explore

  • News
  • Technology
  • AI
  • Science
  • World

Useful Links

  • About Us
  • Contact Us
  • Fact Checking Policy
  • Terms & Conditions
  • Privacy Policy
  • Copyright Policy

Subscribe Us

Subscribe to our newsletter for the Latest News and Top Stories!

© 2022 VellaTimes • All Rights Reserved.
  • Home
  • Web Stories
  • Bookmarks
  • Interests
  • Disclaimer
  • Sitemap
adbanner
AdBlocker Detected
Our site is an advertising supported site. Please whitelist us to support our work.
Okay, I'll Whitelist