Exclusive: Google names new chief of AI infrastructure buildout

With the rapid buildout of infrastructure taking center stage in the tight AI race between tech giants, Google is tapping one of its veteran leaders for a new role overseeing the complex strategy, according to an internal memo reviewed by Semafor.Amin Vahdat, who joined Google from academia roughly 15 years ago, will be named chief technologist for AI infrastructure, according to the memo, and become one of 15 to 20 people reporting directly to CEO Sundar Pichai.Google estimates it will have spent more than $90 billion on capital expenditures by the end of 2025, most of it going into the part of the company Vahdat will now oversee.“This change establishes AI Infrastructure as a key focus area for the company,” wrote Google Cloud CEO Thomas Kurian in the Wednesday memo congratulating Vahdat.Google has been on a roll lately, with high praise for its latest Gemini 3 AI model that has spooked even OpenAI, whose CEO recently declared Google’s ascendance as an emergency for the maker of ChatGPT.But Google’s advantage over OpenAI isn’t just its AI research capabilities: It’s the ability to serve its AI-enabled products to a massive user base efficiently.That capability stems from a decade-long pursuit of custom AI chips, which it calls Tensor Processing Units, and all of the software and equipment built around them to maximize performance and at the lowest possible cost.The Google DeepMind team works with the TPU team to optimize chips for Gemini’s capabilities, and the company has built everything from optical circuit switches to liquid cooling around them in a vertically integrated technology and software stack.Vahdat has been at the center of that effort for years, according to people familiar with his role at the company, long before the AI race put it under the microscope of everyone from tech journalists to Wall Street analysts.In a 2022 blog post that got little attention at the time, Vahdat laid out how his team had transformed the company’s Jupiter network, the system for connecting everything inside data centers.The result was reduced costs for serving core products like YouTube, Search, and Cloud. But it would also prove to be key in the upcoming AI race, where data center interconnects have become a key bottleneck. Frontier models require huge amounts of data to be passed back and forth between processors, allowing a cluster of hundreds of thousands of computers to act as one.Vahdat has also overseen the company’s Borg initiative, a software system that orchestrates a massive number of data center operations, another critical element in today’s AI race. The ability to shuffle an impossibly large number of tasks to ensure every drop of usable compute power is available is crucial to making the economics of AI data centers work. In the AI age, it’s like a game of Tetris but on a galactic scale.In August, Google disclosed in a paper co-authored by Vahdat that the amount of energy used to run the median prompt on its AI models was equivalent to watching less than nine seconds of television and consuming five drops of water. The numbers were far less than what some critics had feared and competitors had likely hoped for.There’s no single answer for how to best run an AI data center. It’s small, coordinated efforts across disparate teams that span the globe. The job of coordinating it all now has an official title.
AI Article