Google Cloud on Wednesday introduced that its eighth era of custom-built AI chips, or tensor processing items (TPUs), shall be cut up in two. One chip, named the TPU 8t, shall be geared for mannequin coaching and one other, the TPU 8i, is geared toward inference.
Inference is the continuing utilization of fashions, aka what occurs after customers submit prompts.
As you would possibly count on, the corporate touts some spectacular efficiency specs for these new TPUs in comparison with the earlier generations: as much as 3x quicker AI mannequin coaching, 80% higher efficiency per greenback, and the power to get 1 million+ TPUs to work collectively in a single cluster. The upshot needs to be much more compute for lots much less vitality — and price to clients — than earlier variations. It calls these chips TPUs, not GPUs, as a result of its {custom} low-power chips had been initially named Tensor.
However Google’s chips aren’t a full frontal assault on Nvidia’s future, at the least not but. Like the opposite big cloud suppliers, together with Microsoft and Amazon, Google is utilizing these chips to complement the Nvidia-based methods it provides in its infrastructure. It’s not flat-out changing Nvidia. In actual fact, Google guarantees its cloud can have Nvidia’s newest chip, Vera Rubin, obtainable later this 12 months.
In the future the hyperscalers constructing their very own AI chips (which incorporates Amazon, Microsoft, and Google) could develop to wish Nvidia much less, as enterprises transfer their AI must their clouds and port their apps to those chips.
Nonetheless, as issues stand at this time, it’s not worthwhile to guess towards Nvidia. As notable chip market analyst Patrick Moorhead jokingly posted on X, he had predicted that Google’s TPU might be dangerous information for Nvidia (and Intel) again in 2016 when the search big launched its first one. Nvidia is now a virtually $5 trillion market cap firm, which means that prediction didn’t precisely maintain as much as the take a look at of time.
If all goes in line with Nvidia’s plan, Google’s development as an AI cloud supplier would end in extra enterprise for the chip maker not much less, even when many a workload runs on Google’s chips.
Techcrunch occasion
San Francisco, CA
|
October 13-15, 2026
In actual fact, Google additionally says it has agreed to work with Nvidia to engineer pc networking that permits Nvidia-based methods to carry out much more effectively in its cloud. Specifically, the 2 tech giants are working to beef up the software-based networking tech known as Falcon, which Google created and open sourced in 2023 underneath the godfather of all open supply knowledge heart {hardware} organizations, the Open Compute Mission.
While you buy by hyperlinks in our articles, we could earn a small fee. This doesn’t have an effect on our editorial independence.





