Google unveils 6th generation TPU 'Trillium', supporting Google Cloud AI with 4.7x better performance and 67% more energy efficiency per chip than TPU v5e



During

the Google I/O 2024 keynote, Google announced its sixth-generation Tensor Processing Unit (TPU), Trillium .

Introducing Trillium, sixth-generation TPUs | Google Cloud Blog
https://cloud.google.com/blog/products/compute/introducing-trillium-6th-gen-tpus/






Google has announced its sixth-generation TPU, Trillium, which is the most performant and energy-efficient to date. Trillium delivers 4.7x better peak performance per chip than the previous-generation TPU v5e . To achieve this, Google is expanding the size and increasing the clock speed of the TPU's matrix multiplication unit (MXU) .



Trillium also doubles the capacity and bandwidth of its high-bandwidth memory (HBM) compared to the TPU v5e, allowing it to operate on larger models with larger key-value caches. The next-generation HBM adopted by Trillium enables high memory bandwidth, improved power efficiency, and a flexible channel architecture, and also improves memory throughput. This reduces the training time and processing latency of large models.

Additionally, Trillium has doubled the inter-chip interconnect (ICI) bandwidth compared to TPU v5e. The doubled ICI bandwidth enables the strategic combination of a custom optical ICI interconnect with 256 chips in a pod and

Google Jupiter Networking to scale to hundreds of pods in a cluster, allowing training and inference jobs to scale to tens of thousands of chips.

In addition, it is equipped with the third-generation SparseCore, a specialized accelerator for processing ultra-large-scale embeddings common in advanced ranking and recommendation workloads. SparseCore is said to be successful in accelerating embedding-heavy workloads by strategically offloading random, fine-grained access from TensorCore.

Trillium will enable us to train the next wave of foundational models faster, reduce latency and lower cost to serve those models, and is over 67% more energy efficient than the TPU v5e.

Trillium can scale up to 256 chips in a single high-bandwidth, low-latency pod. Beyond this pod-level scalability, using multi-slice technology and Titanium Intelligence Processing Units (IPUs), Trillium can scale to hundreds of pods, connecting tens of thousands of chips in building-scale supercomputers interconnected at multi-petabits per second.

Trillium will be available to Google Cloud users in the second half of 2024.



The Google I/O 2024 keynote can be viewed below, and you can check out Trillium's presentation from around the 39 minute mark of the video.

Google Keynote (Google I/O '24) - YouTube


in Hardware,   Video, Posted by logu_ii