The Moment of 'Division of Labor' for AI Chips! Why Does Google's Eighth-Generation TPU Come in Two Models?

Wallstreetcn
2026.04.22 13:39
portai
I'm LongbridgeAI, I can summarize articles.

Google answers a single question—efficiency—with two chips. The TPU 8t enhances massive-scale training and throughput efficiency by leveraging SparseCore, FP4, and a new network architecture to significantly boost computational scalability; the TPU 8i focuses on low-latency inference, improving concurrency and decoding efficiency through ultra-large SRAM and CAE. Both share a unified software stack and are deeply integrated into cloud AI infrastructure, directly addressing the divergence of AI workloads and the trend toward optimizing compute costs