Google Launches Separate AI Chips for Training and Inference, Boosting Performance 2.8x

Gate News message, April 23 — Google announced on April 22 that it will release separate eighth-generation TPU chips for training and inference later this year, replacing its previous combined design. The move targets AI agent workloads and offers Google Cloud customers an alternative to Nvidia hardware.

The training chip delivers 2.8 times the performance of Google’s seventh-generation Ironwood TPU at the same price, while the inference chip is 80% faster and features 384 MB of SRAM, triple the amount in Ironwood. The separation of training and inference capabilities reflects a shift in how companies optimize for different computational demands.

The initiative is backed by a long-term partnership with Broadcom and Anthropic. Anthropic plans to use approximately 3.5 gigawatts of TPU computing through Broadcom starting in 2027, with Broadcom handling chip manufacturing and networking components through 2031. Anthropic, the AI startup behind Claude, has seen annualized revenue recently exceed $30 billion. Meanwhile, Apple, Microsoft, Meta, and Amazon are also expanding custom AI chip efforts to reduce reliance on Nvidia, which remains the market leader.

Disclaimer: The information on this page may come from third parties and does not represent the views or opinions of Gate. The content displayed on this page is for reference only and does not constitute any financial, investment, or legal advice. Gate does not guarantee the accuracy or completeness of the information and shall not be liable for any losses arising from the use of this information. Virtual asset investments carry high risks and are subject to significant price volatility. You may lose all of your invested principal. Please fully understand the relevant risks and make prudent decisions based on your own financial situation and risk tolerance. For details, please refer to Disclaimer.

Related Articles

Taiwan banks team up to build local AI! Finance’s large language model goes live by the end of the year at the fastest

CITIC Financial Holding, led by CITIC Financial Holding’s 16 financial institutions, announced the launch of the “Financial Large Language Model FinLLM” project. The first release of the banking model is expected to be published in August, and in 2026 Q1, AI agents based on FinLLM will be introduced. Training will begin in May, with a budget of approximately 40–70 million yuan. Due to regulatory and localization needs, local data training will be the core, strengthening sovereign AI, building shared infrastructure, and extending to inclusive finance. The plan has been incorporated into the national AI development plan and has received cross-ministry support.

ChainNewsAbmedia16m ago

Google CEO: Capital expenditures in 2026 will reach $185 billion; ramping up investment in the era of AI agents

Google CEO Sundar Pichai announced at Google Cloud Next in Las Vegas on April 22 that Google plans to invest $175 billion to $185 billion in 2026 in capital expenditures to build the infrastructure needed for autonomous AI agents, up from $31 billion in 2022.

MarketWhisper51m ago

Google Jules releases a new version candidate list, repositioning it as an end-to-end product development platform

According to the official April 23 announcement by the Google Jules team, Jules’s product positioning has been upgraded from an asynchronous coding agent to an “end-to-end agentic product development platform.” The new version can read the full product context, independently determine the next steps for building, and submit a PR. The official also announced that the new version candidate list is now open.

MarketWhisper57m ago

Google Jules Rebrands as End-to-End Agentic Product Development Platform, Opens Waitlist for New Version

Gate News message, April 23 — Google's Jules team announced the opening of a waitlist for a new version of the product, repositioning Jules from an asynchronous coding agent to an end-to-end agentic product development platform. According to the official description, the upgraded platform reads enti

GateNews1h ago

Perplexity Discloses Web Search Agent Post-Training Method; Qwen3.5-Based Model Outperforms GPT-5.4 on Accuracy and Cost

Perplexity uses SFT followed by RL with Qwen3.5 models, leveraging a multi-hop QA dataset and rubric checks to boost search accuracy and efficiency, achieving best-in-class FRAMES performance. Abstract: Perplexity's post-training workflow for web-search agents combines supervised fine-tuning (SFT) to enforce instruction-following and language consistency with online reinforcement learning (RL) via the GRPO algorithm. The RL stage uses a proprietary multi-hop verifiable QA dataset and rubric-based conversational data to prevent SFT drift, with reward gating and within-group efficiency penalties. Evaluation shows Qwen3.5-397B-SFT-RL achieving top FRAMES performance, 57.3% accuracy with a single tool call and 73.9% with four calls at $0.02 per query, outperforming GPT-5.4 and Claude Sonnet 4.6 on these metrics. Pricing is API-based and excludes caching.

GateNews2h ago

TikTok Removes Over 538,000 AI-Generated Unauthorized Videos; Multiple Platforms Launch Governance Initiatives

Gate News message, April 23 — TikTok announced a comprehensive crackdown on AI-generated content that infringes on user rights, disclosing that it has removed over 538,000 videos and penalized more than 4,000 accounts to date. The platform will prioritize enforcement against AI deepfakes, voice

GateNews2h ago
Comment
0/400
No comments