Meta Unveils Four Custom MTIA AI Chips Targeting 2027 Deployment



James Ding
Mar 11, 2026 18:12

Meta accelerates custom silicon strategy with MTIA 300-500 chips, promising 4.5x bandwidth gains and 25x compute improvements for GenAI inference by 2027.





Meta dropped details on four generations of custom AI accelerators Wednesday, marking an aggressive push to reduce dependence on Nvidia while serving billions of daily AI interactions across its platforms.

The Meta Training and Inference Accelerator (MTIA) family now spans chips numbered 300 through 500, with the company claiming it can ship new silicon roughly every six months. MTIA 300 is already running in production for ranking and recommendation training, while the 400, 450, and 500 variants target mass deployment through 2027.

The Numbers That Matter

From MTIA 300 to 500, Meta’s claiming a 4.5x increase in high-bandwidth memory throughput and a 25x jump in compute FLOPS when comparing MX8 to MX4 precision formats. The MTIA 450 specifically doubles HBM bandwidth versus the 400, while the 500 adds another 50% on top of that.

For context: HBM bandwidth is the bottleneck for large language model inference. More bandwidth means faster token generation, which translates directly to cost savings at Meta’s scale.

The 400-series delivers 400% higher FP8 FLOPS and 51% higher HBM bandwidth compared to the 300. A single rack houses 72 MTIA 400 devices forming one scale-up domain—competitive positioning against commercial alternatives, according to Meta.

Why Build Custom Silicon?

This announcement came just weeks after Meta signed massive deals with Nvidia and AMD, so the company isn’t abandoning GPU vendors. The strategy is portfolio diversification.

“Mainstream GPUs are typically built for the most demanding workload—large-scale GenAI pre-training—and then applied, often less cost-effectively, to other workloads,” Meta’s engineering team wrote. MTIA flips that approach, optimizing first for inference then adapting elsewhere.

The modular chiplet design allows Meta to swap components without full redesigns. MTIA 400, 450, and 500 share identical chassis, rack, and network infrastructure—new chips drop into existing data center footprints.

The Spending Context

Meta’s infrastructure appetite is staggering. CEO Mark Zuckerberg indicated plans to spend “at least $600 billion” on U.S. data centers and infrastructure through 2028, according to September 2025 reports. Capital expenditure projections for 2025 alone ranged from $60 billion to $65 billion.

Custom silicon doesn’t replace that spending—it optimizes it. Better price-per-performance on inference workloads could meaningfully impact operating costs when you’re running AI recommendations for 3+ billion daily users.

Technical Architecture

Each MTIA chip combines compute chiplets, network chiplets, and HBM stacks. The processing elements contain dual RISC-V vector cores, dedicated engines for matrix multiplication and reductions, plus DMA controllers for memory management.

The software stack runs PyTorch-native, integrating with torch.compile and vLLM’s plugin architecture. Meta claims models can deploy simultaneously on GPUs and MTIA without rewrites—friction reduction that matters for engineering velocity.

MTIA 450 deployment begins early 2027, with the 500 following later that year. Whether these chips deliver on Meta’s performance claims at production scale remains the open question worth watching.

Image source: Shutterstock


Share with your friends!

Products You May Like

Leave a Reply

Your email address will not be published. Required fields are marked *

Please enter CoinGecko Free Api Key to get this plugin works.