Google Ironwood TPU: 10x performance + four partners taking on Nvidia

ChainNewsAbmedia

According to a deep report by Bloomberg and an official announcement from Google, on April 22 Google officially expanded its in-house AI chip lineup: the inference-focused Ironwood (7th-generation TPU) is now fully available on Google Cloud, and it is also simultaneously launching next-generation design collaborations with four partners—Broadcom, MediaTek, Marvell, and Intel. The goal is to use a custom chip supply chain to directly challenge Nvidia’s dominant position in the AI compute market.

Ironwood: 7th-generation TPU, first designed specifically for inference

Ironwood is the 7th-generation product in Google’s TPU series and the first chip designed specifically for inference under the company’s strategy of “training and inference separation.” The specifications Google disclosed are as follows: peak performance per chip is 10x that of TPU v5p, with 192GB of HBM3E memory, memory bandwidth of 7.2 TB/s, and each individual superpod can scale to 9,216 liquid-cooled Ironwood units. Total FP8 compute reaches 42.5 exaflops.

Google’s official statement says Ironwood has been “fully opened for Google Cloud customers to use,” and this year shipment volume is expected to reach a “million-unit” level. Anthropic has committed to using up to 1 million Ironwood TPUs. Meta has signed a “multi-year contract worth billions of dollars” to use TPUs via Google Cloud.

Four-partner division of labor: training goes to Broadcom, inference goes to MediaTek

Google’s next-generation chip supply chain is clearly divided as follows:

Partner Code Role Highlights Broadcom Sunfish Training-focused Dedicated Continue the existing TPU collaboration relationship; leads large training nodes MediaTek (聯發科) Zebrafish Inference-focused Claimed to be 20–30% lower cost than the Broadcom solution Marvell In negotiations Memory Processing Unit (MPU) + additional inference TPU, optimized for HBM and inference Intel Not disclosed Involved in design, strengthening supply-chain diversification

This is the first “four-partner parallel, with training and inference clearly split” model seen in the AI industry. By dispersing IP risk and negotiating power, Google avoids the structural dependency of Nvidia as a single-supplier setup. The roadmap extends to TPU v8 by the end of 2027, to be produced using TSMC’s 2nm process.

Strategic significance: what Google is challenging isn’t a single chip, but the supply chain

Over the past three years, Nvidia has nearly monopolized the AI chip market, with CUDA software ecosystems and the H100/GB200 chips forming a double moat. Google’s Ironwood and four-partner strategy are not aimed at “surpassing at a single point in specs.” Instead, the goal is to replicate Nvidia’s logic in the industry—a “standardized platform + multi-customer procurement” model—so that TPUs are not just for Google’s own use, but a commercial compute option that major AI companies like Anthropic and Meta can share.

The significance of Anthropic’s commitment to 1 million TPUs is especially critical: this is the largest compute commitment from any single AI company outside Nvidia. It also complements Anthropic’s 5GW/100-billion AWS commitment with Amazon reached through AWS on 4/20—one side ties to AWS Trainium, the other ties to Google TPU. With Anthropic’s “dual custom-chip” strategy, Nvidia dependency is reduced. Meta, meanwhile, is the first to publicly include TPUs in its own AI training/inference workloads, sending another signal.

Market reaction and industry linkage

Before this disclosure, MediaTek had already been seen as a “beneficiary of Google’s custom chips.” The exposure of the Zebrafish codename is the first time MediaTek is directly listed as a design partner for Google’s inference chips. This extends the narrative line of “non-Nvidia chip alliances,” including the recent AMD × GlobalFoundries silicon photonics and Marvell × Google MPU initiatives.

Although Nvidia still has GB200 and its next-generation Rubin platform supporting the same period, the compute mix on the customer side is shifting from “all Nvidia” toward “Nvidia + TPU + AWS Trainium” operating on three tracks in parallel. This also means that for TSMC’s 2nm capacity, the four major customers—Google, Nvidia, Apple, and Amazon—are lining up, and negotiating power around foundry supply is continuing to rise.

This article, “Google Ironwood TPU: 10x performance + four-partner lineup to take on Nvidia,” first appeared on Lianxin ABMedia.

Disclaimer: The information on this page may come from third parties and does not represent the views or opinions of Gate. The content displayed on this page is for reference only and does not constitute any financial, investment, or legal advice. Gate does not guarantee the accuracy or completeness of the information and shall not be liable for any losses arising from the use of this information. Virtual asset investments carry high risks and are subject to significant price volatility. You may lose all of your invested principal. Please fully understand the relevant risks and make prudent decisions based on your own financial situation and risk tolerance. For details, please refer to Disclaimer.

Related Articles

SpaceX, OpenAI, and Anthropic IPOs Could Attract Over $240 Billion, Potentially Impacting Crypto Market Liquidity

Gate News message, April 25 — According to market reports, SpaceX is expected to go public in June with a fundraising target exceeding Saudi Aramco's record $29 billion IPO, while OpenAI and Anthropic are planning to list in the second half of 2026. The three companies are projected to

GateNews53m ago

DeepSeek-V4 Preview Released with 1M Context Window, Huawei DCS AI Solution Provides Full Support

Gate News message, April 25 — DeepSeek-V4 preview version has been officially released and open-sourced, featuring an expanded context window of 1 million tokens and introducing KV Cache sliding window with

GateNews1h ago

GPU Shortage Returns as Microsoft, Amazon Tighten Supply; AI Startups Face 32% Price Hike and Year-End Queues

Gate News message, April 25 — A GPU shortage is resurfacing as major cloud providers including Microsoft and Amazon concentrate computing capacity toward internal teams and major customers like OpenAI and Anthropic, leaving smaller AI startups facing price increases, extended wait times, and stricte

GateNews3h ago

Nvidia Deploys OpenAI Codex AI Agent Across Entire Workforce on Blackwell Infrastructure

Gate News message, April 25 — Nvidia has rolled out OpenAI's Codex, an AI agent powered by GPT-5.5, to its entire workforce following a successful trial with approximately 10,000 employees, according to internal communications from CEO Jensen Huang and OpenAI CEO Sam Altman. Codex is designed to as

GateNews3h ago

US State Dept Warns on DeepSeek AI Model Distillation

The US State Department issued a cable on April 24 to diplomatic and consular posts worldwide warning foreign governments about Chinese efforts to copy American AI systems through distillation, according to Reuters. The cable

CryptoFrontier4h ago

Stanford Professor's Health AI Startup Seeks $100M at $1B Valuation

Gate News message, April 25 — Stanford professor James Zou is raising approximately $100 million for Human Intelligence, a California startup developing AI models for human physiology, at a valuation of roughly $1 billion. The company builds on Zou's research in physiology and AI. His lab's

GateNews4h ago
Comment
0/400
No comments