In this exclusive Bloomberg interview, AWS CEO Matt Garman reflects on his first year leading Amazon Web Services, the cloud computing giant driving much of Amazon’s growth. Garman discusses AWS’s aggressive global expansion, including recent data center developments in Mexico, Chile, New Zealand, Saudi Arabia, and Taiwan, as part of its strategy to meet surging AI infrastructure demands.
Garman explains how AWS is scaling up to support artificial intelligence workloads, including training and inference, by collaborating with chipmakers like Nvidia and deploying custom silicon such as AWS’s Trainium2. He emphasizes the increasing role of AI across all Amazon operations and customer solutions, highlighting a multibillion-dollar run rate for generative AI services.
He also shares insights into AWS’s collaboration with AI model developers like Anthropic and expresses openness to hosting OpenAI models on AWS. Garman touches on customer demand for flexibility and performance, the rise of inference as a core application function, and how token-based metrics are evolving in AI workloads.
The conversation offers a comprehensive look at AWS’s role in the rapidly evolving cloud and AI landscape, underscoring Garman’s focus on customer choice, infrastructure scale, and technical leadership in the next wave of digital transformation.