The pace of technological change since last year’s Cloud Next has never been faster. Our first-party models now process more than 16 billion tokens per minute via direct API use, up from 10 billion last quarter. To support this growth, in 2026, just over half of our overall machine learning compute investment is expected to go towards the Cloud business to benefit our customers and partners.
This week at Cloud Next we’re sharing an extraordinary range of new partnerships and innovations, including our new Gemini Enterprise Agent Platform, the new mission control to build, scale, govern, and optimize agents.
We also unveiled our eighth-generation TPUs, including TPU 8t for training and TPU 8i for inference, to provide the massive throughput needed to run millions of agents (great to get my hands on these!)
To be the best partner, we always want to be “customer zero” for our own technologies. This helps us imagine, test, build and scale the best Google technologies for our cloud customers, for today and tomorrow. Our database service Bigtable, which powers so many Google services, and our TPUs, which have been so important in training and powering our Gemini models, are great examples. We’ve been using AI to generate code internally at Google for a while. Today, 75% of all new code at Google is now AI-generated and approved by engineers, up from 50% last fall. And teams across the company — from marketing to security — are using AI to work more effectively.
We’re firmly in our agentic Gemini era. Read more of what we announced at Cloud Next here:
https://lnkd.in/e8JXGbqV