Topics

a la mode

AI

Amazon

Article image

Image Credits:Credit: kynny / Getty Images

Apps

Biotech & Health

Climate

Article image

Image Credits:Google

Cloud Computing

commercialism

Crypto

Article image

Image Credits:Google

Enterprise

EVs

Fintech

Fundraising

Gadgets

Gaming

Google

Government & Policy

ironware

Instagram

layoff

Media & Entertainment

Meta

Microsoft

Privacy

Robotics

Security

Social

distance

startup

TikTok

Transportation

Venture

More from TechCrunch

event

Startup Battlefield

StrictlyVC

Podcasts

video

Partner Content

TechCrunch Brand Studio

Crunchboard

Contact Us

Google today announced the launch of its newGeminilarge language model ( LLM ) and with that , the company also establish its new Cloud TPU v5p , an update version of its Cloud TPU v5e , whichlaunchedinto general availability earlier this twelvemonth . A v5p pod consists of a sum of 8,960 chips and is second by Google ’s fast interconnect yet , with up to 4,800 Gpbs per chip .

It ’s no surprisal that Google assure that these chip are significantly faster than the v4 TPUs . The team claims that the v5p features a 2x improvement in floating-point operation and 3x improvement in high - bandwidth memory . That ’s a mo like comparing the new Gemini model to the aged OpenAI GPT 3.5 example , though . Google itself , after all , already moved the state of the graphics beyond the TPU v4 . In many ways , though , v5e pods were a fleck of a downgrade from the v4 pod , with only 256 v5e chips per pod versus 4096 in the v4 seedpod and a totality of 197 TFLOPs 16 - bit floating point in time performance per v5e chip versus 275 for the v4 crisp . For the new v5p , Google promises up to 459 TFLOPs of 16 - bit floating point carrying into action , back by the fast interconnect .

Google says all of this entail the TPU v5p can train a large oral communication model like GPT3 - 175B 2.8 times faster than the TPU v4 — and do so more cost - in effect , too ( though the TPU v5e , while slower , really extend more comparative functioning per dollar than the v5p ) .

“ In our early stage usage , Google DeepMind and Google Research have keep 2X speedups for LLM training workloads using TPU v5p chips liken to the performance on our TPU v4 generation , ” pen Jeff Dean , main scientist , Google DeepMind and Google Research . “ The racy bread and butter for ML Frameworks ( JAX , PyTorch , TensorFlow ) and orchestration tools enables us to surmount even more expeditiously on v5p . With the 2nd genesis of SparseCores we also see significant betterment in the performance of embeddings - hard workloads . TPUs are vital to enable our largest - scale research and engineering effort on cutting edge models like Gemini . ”

The new TPU v5p is n’t generally useable yet , so developers will have to reach out to their Google account director to get on the listing .