Google unveils two new TPUs designed for the “agentic era”


So the new chips allow for faster training, but Google also says you get more useful computation for every volt you pump into a TPU 8t. The company claims a “goodpute” rate of 97 percent, which means less waiting and wasted effort. With better handling of irregular memory access, automatic handling of hardware faults, and real-time telemetry across all connected chips, TPU 8t spends more time actively advancing model training.

When training is done, AI models run in inference mode to generate tokens—that’s the process happening behind the scenes when you tell a model to do something. This doesn’t require as much horsepower, so using the same hardware for both parts of the AI lifecycle is inefficient. That’s why inference is the purview of TPU 8i, which is designed to be more efficient when running multiple specialized agents, with less waiting time. TPU 8i chips also run in larger pods of 1,152 chips versus just 256 for the last-gen Ironwood inference clusters. That works out to 11.6 EFlops per pod, much lower than TPU 8t pods.

TPU 8i chip

The TPU 8i has less raw power than TPU 8t.

Credit:
Google

The TPU 8i has less raw power than TPU 8t.


Credit:

Google

Google has tripled the amount of on-chip SRAM for each TPU 8i to 384 MB. This allows the company’s new chips to keep a larger key value cache on the chip, speeding up models with longer context windows. The eighth-gen AI accelerators are also the first from Google to rely solely on Google’s custom Axion ARM CPU host, featuring one CPU for every two TPUs. In Ironwood, each x86 CPU serviced four TPU chips. Google says this “full-stack” ARM-based approach allows for much greater efficiency.

An efficiency play

It makes sense that efficiency is a core part of Google’s new TPU setup. Training and running frontier AI models is expensive, and the return on investment is unclear. Companies are still burning money on generative AI in the hopes that efficiency will turn the corner at some point. Maybe Google’s new TPUs will help get there and maybe not, but the company has made notable improvements.



Source link

  • Related Posts

    LinkedIn’s CEO is moving on; please hold your tearful video tributes

    Ryan Roslansky has stepped down as LinkedIn’s CEO after six years running the world’s largest professional network. Dan Shapero, the company’s COO, takes over immediately. Source link

    Anthropic investigates report of rogue access to hack-enabling Mythos AI | AI (artificial intelligence)

    The AI developer Anthropic has confirmed it is investigating a report that unauthorised users have gained access to its Mythos model, which it has warned poses risks to cybersecurity. The…

    Leave a Reply

    Your email address will not be published. Required fields are marked *

    You Missed

    Trump now says Iran’s government is ‘seriously fractured’ after previously praising new leaders

    Trump now says Iran’s government is ‘seriously fractured’ after previously praising new leaders

    Canadian immigration minister's sporadic communication about new permanent resident program fuelling misinformation, lawyers charge

    Canadian immigration minister's sporadic communication about new permanent resident program fuelling misinformation, lawyers charge

    Former MrBeast staffer accuses YouTuber’s company of sexism and leave violations

    Former MrBeast staffer accuses YouTuber’s company of sexism and leave violations

    LinkedIn’s CEO is moving on; please hold your tearful video tributes

    LinkedIn’s CEO is moving on; please hold your tearful video tributes

    Ecco the Dolphin: Complete revealed, a mixture of remaster and brand new game spanning every Ecco game from the 8 to 16-bit era

    Ecco the Dolphin: Complete revealed, a mixture of remaster and brand new game spanning every Ecco game from the 8 to 16-bit era

    Why The US Air Force Still Relies On The Boeing 707 In 2026

    Why The US Air Force Still Relies On The Boeing 707 In 2026