Google has introduced the sixth generation of its custom hardware units, known as Trillium, at the highly anticipated Google I/O 2024 event. The Tensor Processing Units (TPUs)—specially designed chips—serve as the backbone of the company's cutting-edge cloud hypercomputing solutions aimed at enterprises.
This latest TPU generation represents a groundbreaking leap, delivering 4.7 times the peak compute performance per chip compared to the previous TPU v5e model. In addition to enhanced performance, Trillium TPUs feature double the High Bandwidth Memory (HBM) capacity, double the bandwidth, and double the Interchip Interconnect (ICI) bandwidth relative to their predecessors. Notably, this new generation integrates the third iteration of SparseCore, an advanced accelerator specifically developed for handling ultra-large embeddings, a critical component for powering sophisticated ranking and recommendation systems.
Designed with efficiency in mind, the Trillium chips expedite AI training workloads, achieving faster processing with reduced latency and lower operational costs. Remarkably, they are 67% more energy-efficient than their TPU v5e counterparts. The Trillium pods will facilitate scalability for cloud customers, accommodating up to 256 TPUs to meet diverse computing demands.
Sundar Pichai, Google’s CEO, emphasized the growing need for robust computing capabilities. "Training state-of-the-art models requires substantial computing power. The industry demand for machine learning compute has surged by a factor of one million over the past six years, and this demand only continues to grow, increasing tenfold each year. Google was designed to meet these challenges. For 25 years, we have invested heavily in developing world-class technical infrastructure, from the cutting-edge hardware that supports our search engine to our custom TPUs that drive our AI advancements."
The Trillium TPUs are set to be available to Google Cloud customers in late 2024, paving the way for more accessible and efficient AI services.
Dr. Andrew Rogoyski from the University of Surrey's Institute for People-Centred AI hailed Google's announcement as a significant milestone. "The latest TPU generation will be a critical enabler for providing affordable AI services in the future. Increasingly, AI companies are focusing on developing AI hardware as a sustainable path forward in the industry. However, chip design and manufacturing remain domains that only the most financially robust firms can afford to explore."
In addition to the Trillium launch, Google shared updates regarding its support for Nvidia’s new Blackwell hardware. CEO Pichai announced that Google Cloud will be among the first cloud providers to roll out support for Blackwell in early 2025. "We are fortunate to have a longstanding partnership with Nvidia and are excited to bring Blackwell's capabilities to our customers," Pichai stated.
With these advancements, Google continues to position itself at the forefront of AI innovation, responding to the growing demands of the industry while enhancing the capabilities available to developers and enterprises worldwide.