google cloud

Google Cloud strengthens its AI infrastructure with TPU Ironwood and Cloud WAN

Launch of Google Cloud Next 2025 in Las Vegas

The Google Cloud Next 2025 event kicked off in Las Vegas from April 9 to 11, drawing nearly 32,000 attendees for this major gathering. Sundar Pichai, CEO of Alphabet, opened the conference by emphasizing the growing importance of high-performance, low-latency infrastructure to support advancements in artificial intelligence.

In this context, he announced massive investments of $75 billion in the development of new data centers, aiming to counter competing initiatives such as Stargate (led by OpenAI, Oracle, and SoftBank) and to meet the increasing demand for AI computing power.

Ironwood: The New Generation of TPUs for Inference

Google Cloud has introduced Ironwood, the seventh generation of its Tensor Processing Units (TPUs), specifically designed for inference tasks—operations that generate outputs from pre-trained AI models. This chip is the successor to the Trillium model, released in 2024.

Key features of Ironwood:

  • Available in pods ranging from 256 to 9,126 chips

  • Up to 45.2 exaflops of compute power

  • Each chip delivers 4.16 teraflops

  • Up to 192 GB of HBM memory per pod, with a bandwidth of 7.2 TB/s

  • Enhanced chip-to-chip communication using ICI (Inter-Chip Interconnect) at 1.2 TB/s

Ironwood also integrates SparseCore, a specialized accelerator for large-scale embeddings, along with Pathway, a TPU-distributed execution environment developed by DeepMind.

Enhanced AI Instances with Nvidia Blackwell Chips

Although Google is highlighting its own chips, Nvidia remains a key strategic partner. Google Cloud now supports:

  • A4 instances powered by Blackwell B200 accelerators

  • New A4X instances (currently in preview), built on the GB200 NVL72 architecture, which combines Grace CPUs (Arm-based) with Blackwell GPUs

The infrastructure also benefits from Cluster Director, a tool designed to manage clusters of AI accelerators as a single, optimized unit.

Additionally, Google Cloud has announced it will be the first provider to offer Nvidia’s Rubin GPUs, which were recently unveiled at GTC.

Cloud WAN: A High-Performance Global Network for AI

To support its AI offerings, Google introduces Cloud WAN, a networking solution built on Google’s global infrastructure:

  • Over 2 million kilometers of fiber optic cable

  • 33 submarine cables

  • 202 points of presence worldwide

Cloud WAN is designed for enterprises with inter-regional or multi-site connectivity needs. It offers:

  • Cloud Interconnect: connects enterprise data centers directly to Google Cloud

  • Cross-Cloud Interconnect: enables multicloud setups with direct links to other cloud providers

  • Cross-Site Interconnect (in preview): provides private, point-to-point connections at speeds of up to 100 Gbps, enabling fast data exchange between different locations

Conclusion: A Bold AI Strategy in the Face of Competition

With the introduction of Ironwood, the expansion of Nvidia-powered offerings, and the strengthening of its network through Cloud WAN, Google Cloud clearly demonstrates its ambition to lead the global AI infrastructure landscape, while gearing up for fierce competition in a rapidly growing sector.

source of the news: LeMondeInformatique