Forgot your password?
typodupeerror
Google AI Hardware

Google Unveils Two New AI Chips For the 'Agentic Era' (cnbc.com) 18

Google announced two new tensor processing units (TPUs) for the "agentic era," with separate processors dedicated to training and inference. "With the rise of AI agents, we determined the community would benefit from chips individually specialized to the needs of training and serving," Amin Vahdat, a Google senior vice president and chief technologist for AI and infrastructure, said in a blog post. Both chips will become available later this year. CNBC reports: After years of producing chips that can both train artificial intelligence models and handle inference work, Google is separating those tasks into distinct processors, its latest effort to take on Nvidia in AI hardware. [...] None of the tech giants are displacing Nvidia, and Google isn't even comparing the performance of its new chips with those from the AI chip leader. Google did say the training chip enables 2.8 times the performance of the seventh-generation Ironwood TPU, announced in November, for the same price, while performance is 80% better for the inference processor.

Nvidia said its upcoming Groq 3 LPU hardware will draw on large quantities of static random-access memory, or SRAM, which is used by Cerebras, an AI chipmaker that filed to go public earlier this month. Google's new inference chip, dubbed TPU 8i, also relies on SRAM. Each chip contains 384 megabytes of SRAM, triple the amount in Ironwood. The architecture is designed "to deliver the massive throughput and low latency needed to concurrently run millions of agents cost-effectively," Sundar Pichai, CEO of Google parent Alphabet, wrote in a blog post.

Google Unveils Two New AI Chips For the 'Agentic Era'

Comments Filter:
  • by Tailhook ( 98486 ) on Wednesday April 22, 2026 @03:08PM (#66107272)

    SRAM has never been built at this scale, afaik. Cerebras was ahead of the curve here, building wafer scale SRAMs years ago. The penalties of DRAM (even with HBM) are now so severe that everyone is taking the gloves off and building mighty SRAMs. This has always been possible in theory, but the high cost never justified it.

    The impact on semiconductor fab demand is significant. SRAM cells are larger than DRAM bits: more silicon die area for the same amount of RAM.

    Also, the training vs. inference split Google is baking into actual hardware is a big deal: it's the reality that training and inference are very distinct things asserting itself, which has been obvious to anyone that hasn't been drinking excessive NVidia cool-aid: there is a future where costly, general purpose GPU-like devices aren't actually necessary for operating LLMs.

    • Also, the training vs. inference split Google is baking into actual hardware is a big deal: it's the reality that training and inference are very distinct things asserting itself, which has been obvious to anyone that hasn't been drinking excessive NVidia cool-aid: there is a future where costly, general purpose GPU-like devices aren't actually necessary for operating LLMs.

      The training/inference requirements are indeed different in significant ways. If the TPU 8i had been available a few years ago, it would have seriously affected Nvidia sales. However, the 8i is just now in the process of becoming available and is also not a small or cheap module. It's also being introduced at the same time that Nvidia is introducing its own inference devices. Due to the market timing and other factors, it remains to be seen how much it affects Nvidia's market.

  • Just wait until the bubble bursts, and everyone starts removing anything 'AI' from their devices (as much as possible), and stops using it because they're sick of the hallucinations or how it's baked-into everything... I'll keep using my Galaxy S9 (and Win10 Enterprise LTSC, and not-smart 18 year old Plasma TV) until it totally dies (only used Bixby like 5 times... mostly just seeing if it's worth using).

    I don't need "Clod" to generate my Arduino code for me... I'll look up stuff on my own. I can type out

  • This is how it goes, the end goal will be that simple users can't do AI training without a permit. You know, for the children.

When you don't know what to do, walk fast and look worried.

Working...