Ironwood is Google’s newest tensor processing unit
Nvidia’s place because the dominant provider of AI chips could also be underneath risk from a specialised chip pioneered by Google, with studies suggesting firms like Meta and Anthropic wish to spend billions on Google’s tensor processing models.
What’s a TPU?
The success of the synthetic intelligence business has been largely based mostly on graphical processing models (GPUs), a type of laptop chip that may carry out many parallel calculations on the similar time, fairly than one after the opposite like the pc processing models (CPUs) that energy most computer systems.
GPUs had been initially developed to help with laptop graphics, because the title suggests, and gaming. “If I’ve a number of pixels in an area and I have to do a rotation of this to calculate a brand new digicam view, that is an operation that may be finished in parallel, for a lot of totally different pixels,” says Francesco Conti on the College of Bologna in Italy.
This capability to do calculations in parallel occurred to be helpful for coaching and working AI fashions, which frequently use calculations involving huge grids of numbers carried out on the similar time, known as matrix multiplication. “GPUs are a really basic structure, however they’re extraordinarily suited to purposes that present a excessive diploma of parallelism,” says Conti.
Nevertheless, as a result of they weren’t initially designed with AI in thoughts, there could be inefficiencies within the ways in which GPUs translate the calculations which can be carried out on the chips. Tensor processing models (TPUs), which had been initially developed by Google in 2016, are as an alternative designed solely round matrix multiplication, says Conti, that are the primary calculations wanted for coaching and working giant AI fashions.
This yr, Google launched the seventh era of its TPU, known as Ironwood, which powers lots of the firm’s AI fashions like Gemini and protein-modelling AlphaFold.
Are TPUs significantly better than GPUs for AI?
Technologically, TPUs are extra of a subset of GPUs than a completely totally different chip, says Simon McIntosh-Smith on the College of Bristol, UK. “They concentrate on the bits that GPUs do extra particularly geared toward coaching and inference for AI, however truly they’re in some methods extra just like GPUs than you may suppose.” However as a result of TPUs are designed with sure AI purposes in thoughts, they are often way more environment friendly for these jobs and save doubtlessly tens or lots of of tens of millions of {dollars}, he says.
Nevertheless, this specialisation additionally has its disadvantages and might make TPUs rigid if the AI fashions change considerably between generations, says Conti. “For those who don’t have the flexibleness in your [TPU], you must do [calculations] on the CPU of your node within the knowledge centre, and this can gradual you down immensely,” says Conti.
One benefit that Nvidia GPUs have historically held is that there’s easy software program accessible that may assist AI designers run their code on Nvidia chips. This didn’t exist in the identical approach for TPUs once they first took place, however the chips at the moment are at a stage the place they’re extra simple to make use of, says Conti. “With the TPU, now you can do the identical [as GPUs],” he says. “Now that you’ve got enabled that, it’s clear that the provision turns into a significant factor.”
Who’s constructing TPUs?
Though Google first launched the TPU, lots of the largest AI firms (referred to as hyperscalers), in addition to smaller start-ups, have now began growing their very own specialised TPUs, together with Amazon, which makes use of its personal Trainium chips to coach its AI fashions.
“A lot of the hyperscalers have their very own inner programmes, and that’s partly as a result of GPUs obtained so costly as a result of the demand was outstripping provide, and it is perhaps cheaper to design and construct your individual,” says McIntosh-Smith.
How will TPUs have an effect on the AI business?
Google has been growing its TPUs for over a decade, however it has principally been utilizing these chips for its personal AI fashions. What seems to be altering now’s that different giant firms, like Meta and Anthropic, are making sizeable purchases of computing energy from Google’s TPUs. “What we haven’t heard about is massive prospects switching, and possibly that’s what’s beginning to occur now,” says McIntosh-Smith. “They’ve matured sufficient and there’s sufficient of them.”
In addition to creating extra selection for the massive firms, it may make good monetary sense for them to diversify, he says. “It’d even be that which means you get a greater deal from Nvidia sooner or later,” he says.
Subjects:

