Google Ironwood TPU: 10x performance + four partners taking on Nvidia

ChainNewsAbmedia

According to a deep report by Bloomberg and an official announcement from Google, on April 22 Google officially expanded its in-house AI chip lineup: the inference-focused Ironwood (7th-generation TPU) is now fully available on Google Cloud, and it is also simultaneously launching next-generation design collaborations with four partners—Broadcom, MediaTek, Marvell, and Intel. The goal is to use a custom chip supply chain to directly challenge Nvidia’s dominant position in the AI compute market.

Ironwood: 7th-generation TPU, first designed specifically for inference

Ironwood is the 7th-generation product in Google’s TPU series and the first chip designed specifically for inference under the company’s strategy of “training and inference separation.” The specifications Google disclosed are as follows: peak performance per chip is 10x that of TPU v5p, with 192GB of HBM3E memory, memory bandwidth of 7.2 TB/s, and each individual superpod can scale to 9,216 liquid-cooled Ironwood units. Total FP8 compute reaches 42.5 exaflops.

Google’s official statement says Ironwood has been “fully opened for Google Cloud customers to use,” and this year shipment volume is expected to reach a “million-unit” level. Anthropic has committed to using up to 1 million Ironwood TPUs. Meta has signed a “multi-year contract worth billions of dollars” to use TPUs via Google Cloud.

Four-partner division of labor: training goes to Broadcom, inference goes to MediaTek

Google’s next-generation chip supply chain is clearly divided as follows:

Partner Code Role Highlights Broadcom Sunfish Training-focused Dedicated Continue the existing TPU collaboration relationship; leads large training nodes MediaTek (聯發科) Zebrafish Inference-focused Claimed to be 20–30% lower cost than the Broadcom solution Marvell In negotiations Memory Processing Unit (MPU) + additional inference TPU, optimized for HBM and inference Intel Not disclosed Involved in design, strengthening supply-chain diversification

This is the first “four-partner parallel, with training and inference clearly split” model seen in the AI industry. By dispersing IP risk and negotiating power, Google avoids the structural dependency of Nvidia as a single-supplier setup. The roadmap extends to TPU v8 by the end of 2027, to be produced using TSMC’s 2nm process.

Strategic significance: what Google is challenging isn’t a single chip, but the supply chain

Over the past three years, Nvidia has nearly monopolized the AI chip market, with CUDA software ecosystems and the H100/GB200 chips forming a double moat. Google’s Ironwood and four-partner strategy are not aimed at “surpassing at a single point in specs.” Instead, the goal is to replicate Nvidia’s logic in the industry—a “standardized platform + multi-customer procurement” model—so that TPUs are not just for Google’s own use, but a commercial compute option that major AI companies like Anthropic and Meta can share.

The significance of Anthropic’s commitment to 1 million TPUs is especially critical: this is the largest compute commitment from any single AI company outside Nvidia. It also complements Anthropic’s 5GW/100-billion AWS commitment with Amazon reached through AWS on 4/20—one side ties to AWS Trainium, the other ties to Google TPU. With Anthropic’s “dual custom-chip” strategy, Nvidia dependency is reduced. Meta, meanwhile, is the first to publicly include TPUs in its own AI training/inference workloads, sending another signal.

Market reaction and industry linkage

Before this disclosure, MediaTek had already been seen as a “beneficiary of Google’s custom chips.” The exposure of the Zebrafish codename is the first time MediaTek is directly listed as a design partner for Google’s inference chips. This extends the narrative line of “non-Nvidia chip alliances,” including the recent AMD × GlobalFoundries silicon photonics and Marvell × Google MPU initiatives.

Although Nvidia still has GB200 and its next-generation Rubin platform supporting the same period, the compute mix on the customer side is shifting from “all Nvidia” toward “Nvidia + TPU + AWS Trainium” operating on three tracks in parallel. This also means that for TSMC’s 2nm capacity, the four major customers—Google, Nvidia, Apple, and Amazon—are lining up, and negotiating power around foundry supply is continuing to rise.

This article, “Google Ironwood TPU: 10x performance + four-partner lineup to take on Nvidia,” first appeared on Lianxin ABMedia.

Disclaimer: The information on this page may come from third parties and does not represent the views or opinions of Gate. The content displayed on this page is for reference only and does not constitute any financial, investment, or legal advice. Gate does not guarantee the accuracy or completeness of the information and shall not be liable for any losses arising from the use of this information. Virtual asset investments carry high risks and are subject to significant price volatility. You may lose all of your invested principal. Please fully understand the relevant risks and make prudent decisions based on your own financial situation and risk tolerance. For details, please refer to Disclaimer.

Related Articles

OpenClaw, Hermes, and SillyTavern Confirmed in GLM Coding Plan Support

Zhipu AI PM Li announces OpenClaw, Hermes, and SillyTavern as supported GLM Coding Plan projects; other tools will be evaluated case-by-case. Do not share credentials or use subscriptions as API access; contact support for error 1313. Zhipu AI product manager Li announced that OpenClaw, Hermes, and SillyTavern are officially supported under the GLM Coding Plan, with other tools evaluated case-by-case. The note cautions against sharing credentials or using subscriptions as API access and directs users with error 1313 to contact support.

GateNews2h ago

Google Cloud CEO: Gemini to Power Apple's Personalized Siri Rollout in 2026

Summary: Gemini will power a personalized Apple Siri in 2026, built on Apple's Foundation Models and Gemini collaboration; Apple tests a chat-like Siri in iOS 27/macOS 27, slated for WWDC 2026. Abstract: Google Cloud's Gemini is set to power a personalized Apple Siri by 2026, blending Gemini with Apple's Foundation Models under a roughly $1 billion collaboration. Apple is testing a redesigned, chat-like Siri in iOS 27/macOS 27, with a Dynamic Island interface and new features, ahead of a WWDC 2026 unveiling on June 8.

GateNews2h ago

SpaceX $60B Cursor Deal Fuels SBF's Pardon Push as FTX's $200K Stake Now Worth $3B

Gate News message, April 22 — SpaceX announced a major partnership with AI coding startup Cursor today, with an option to acquire the company for $60 billion. The deal has given fresh ammunition to Sam Bankman-Fried (SBF), who is currently incarcerated and pushing for a presidential pardon, as it de

GateNews3h ago

Chegg Stock Crashes 99% as AI Disrupts Edtech Market

Summary: Chegg soared during online-education demand, then AI tools disrupted its model, triggering massive layoffs and a collapse below $2, with broader AI-driven shifts hitting crypto miners and fintech firms. Abstract: This article examines Chegg's rise as a pandemic-era edtech darling and its ensuing decline amid the rapid adoption of generative AI, which provides quick answers and undercuts Chegg's value proposition. It documents 2025 layoffs and the stock's plunge toward delisting, and frames Chegg's experience within a broader AI disruption reshaping tech and crypto: Bitcoin miners pivot to AI operations, and AI-native strategies redefine competitiveness in fintech and beyond.

CryptoFrontier3h ago

OpenAI Releases Open-Source Privacy Filter Model for PII Detection and Redaction

Abstract: OpenAI's Privacy Filter is an open-source, locally executable model that detects and redacts PII in text. It supports large contexts, identifies many PII categories, and is intended for privacy-preserving workflows such as data preparation, indexing, logging, and moderation. OpenAI's Privacy Filter is a locally run, open-source model (128k-token context) that detects and redacts PII in text, covering contact, financial, and credential data for privacy workflows.

GateNews3h ago

OpenAI Plans to Deploy 30GW Computing Power by 2030

OpenAI aims for 30GW of computing by 2030 to meet rising AI demands, with 8GW completed of a 10GW 2025 target. The expansion signals a strategy to scale infrastructure for next-generation AI development and deployment. OpenAI intends to reach 30GW of computing power by 2030 to accommodate growing AI demands, having already completed 8GW of a 10GW target for 2025. The move reflects a strategic expansion of infrastructure to support next-generation AI development and deployment.

GateNews3h ago
Comment
0/400
No comments