Google says its AI supercomputer is faster, greener than Nvidia A100 chip

The Google TPU is now in its fourth generation

Alphabet Inc's Google released on Tuesday new details about the supercomputers it uses to train its artificial intelligence models, saying the systems are both faster and more power-efficient than comparable systems from Nvidia Corp.

Google has designed its own custom chip called the Tensor Processing Unit, or TPU. It uses those chips for more than 90% of the company's work on artificial intelligence training, the process of feeding data through models to make them useful at tasks such as responding to queries with human-like text or generating images.

The Google TPU is now in its fourth generation. Google on Tuesday published a scientific paper detailing how it has strung more than 4,000 of the chips together into a supercomputer using its own custom-developed optical switches to help connect individual machines.

The Google logo, on a smartphone, and Bard logo

BRAZIL - 2023/02/09: In this photo illustration, the Google logo is displayed on a smartphone screen, with the page of the new artificial intelligence (AI) tool called Bard in the background.  ((Photo Illustration by Rafael Henrique/SOPA Images/LightRocket via Getty Images) / Getty Images)

Improving these connections has become a key point of competition among companies that build AI supercomputers because so-called large language models that power technologies like Google's Bard or OpenAI's ChatGPT have exploded in size, meaning they are far too large to store on a single chip.

GOOGLE'S AI BARD CAN EASILY BREAK SAFETY FEATURES, PROMOTE COVID AND CLIMATE CONSPIRACIES, STUDY FINDS

The models must instead be split across thousands of chips, which must then work together for weeks or more to train the model. Google's PaLM model - its largest publicly disclosed language model to date - was trained by splitting it across two of the 4,000-chip supercomputers over 50 days.

Ticker Security Last Change Change %
GOOG ALPHABET INC. 123.87 +2.61 +2.15%

Google said its supercomputers make it easy to reconfigure connections between chips on the fly, helping avoid problems and tweak for performance gains.

"Circuit switching makes it easy to route around failed components," Google Fellow Norm Jouppi and Google Distinguished Engineer David Patterson wrote in a blog post about the system. "This flexibility even allows us to change the topology of the supercomputer interconnect to accelerate the performance of an ML (machine learning) model."

Google Bard

Google AI on a mobile phone on Feb. 9, 2023, in Brussels, Belgium.  ((Photo by Jonathan Raa/NurPhoto via Getty Images) / Getty Images)

While Google is only now releasing details about its supercomputer, it has been online inside the company since 2020 in a data center in Mayes County, Oklahoma. Google said that startup Midjourney used the system to train its model, which generates fresh images after being fed a few words of text.

CHATGPT AI LISTS JOBS IT CAN DO BETTER THAN HUMANS AS MILLIONS COULD BE PUT OUT OF WORK

In the paper, Google said that for comparably sized systems, its chips are up to 1.7 times faster and 1.9 times more power-efficient than a system based on Nvidia's A100 chip that was on the market at the same time as the fourth-generation TPU.

The logo of Nvidia Corporation is seen during the annual Computex computer exhibition in Taipei, Taiwan May 30, 2017. REUTERS/Tyrone Siu//File Photo (Reuters Photos)

A Nvidia spokesperson declined to comment.

CLICK HERE TO GET THE FOX BUSINESS APP

Google said it did not compare its fourth-generation to Nvidia's current flagship H100 chip because the H100 came to the market after Google's chip and is made with newer technology.

Ticker Security Last Change Change %
NVDA NVIDIA CORP. 430.25 -0.20 -0.05%

Google hinted that it might be working on a new TPU that would compete with the Nvidia H100 but provided no details, with Jouppi telling Reuters that Google has "a healthy pipeline of future chips."