The pace of technological change since last year’s Cloud Next has never been faster, and Google Cloud has incredible momentum.
Our first-party models now process more than 16 billion tokens per minute via direct API use by our customers, up from 10 billion last quarter. To support and drive this growth, in 2026, just over half of our overall machine learning compute investment is expected to go towards the Cloud business to benefit our cloud customers and partners.
You can read all about our momentum and the extraordinary range of partnerships and innovations we’re announcing at Cloud Next.
I want to highlight just four key areas.
1. We’re firmly in the agentic Gemini era
Last fall we introduced Gemini Enterprise, the end-to-end system for the agentic era — the connective tissue between your data, your people and your goals.
It has great momentum: In Q1, we saw 40% growth in paid monthly active users quarter-over-quarter.
Through this rapid growth, we’ve seen how every employee in every organization can become a builder. This is an incredible shift, but it comes with complexity. The conversation has gone from “Can we build an agent?” to “How do we manage thousands of them?”
That’s why we’re introducing our new Gemini Enterprise Agent Platform. It provides the secure, full-stack connective tissue you need to build, scale, govern and optimize your agents with confidence — a mission control for the agentic enterprise.
2. Using AI to defend against security threats
While AI can increase security risks, our Cloud customers now have AI on their side to protect their organizations. Today we are unveiling a range of new agentic solutions for threat detection, as part of an AI-powered cybersecurity platform that combines Google’s Threat Intelligence and Security Operations with Wiz’s Cloud and AI Security Platform.
In addition, we are launching Wiz’s new AI Application Protection Platform (AI-APP), which provides autonomous protection, from code to cloud to runtime, across multicloud, hybrid and AI environments.
3. Introducing our eighth-generation TPUs
In the era of AI agents, infrastructure needs to evolve to take on the most demanding AI workloads. This year, we’re bringing the eighth generation of our Tensor Processing Units with a dual chip approach:
TPU 8t, optimized for training, scales up to 9,600 TPUs and 2 petabytes of shared, high-bandwidth memory in a single superpod. It achieves three times the processing power of Ironwood and delivers up to 2x more performance/watt.TPU 8i, optimized for inference, connects 1,152 TPUs in a single pod, dramatically reducing latency, with 3x more on-chip SRAM, to deliver the massive throughput and low latency needed to concurrently run millions of agents cost-effectively.We’ll offer these to Cloud customers as a core part of our selection of compute processors, along with a portfolio of NVIDIA GPU instances. Read more in our blog post.
4. Staying on the cutting-edge as “customer zero”
To be the best partner, we always want to be “customer zero” for our own technologies. This helps us imagine, test, build and scale the best Google technologies for our cloud customers, for today and tomorrow. Our database service Bigtable, which powers so many Google services, and our TPUs, which have been so important in training and powering our Gemini models, are great examples.
Here are a few more recent ones:
First, coding.
We’ve been using AI to generate code internally at Google for a while. Today, 75% of all new code at Google is now AI-generated and approved by engineers, up from 50% last fall.We’re now shifting to truly agentic workflows. Our engineers are orchestrating fully autonomous digital task forces, firing off agents and accomplishing incredible things.Recently, a particularly complex code migration done by agents and engineers working together was completed six times faster than was possible a year ago with engineers alone.And with our recent launch of the Gemini app on MacOS, the team built the initial release with our agentic development platform Antigravity, going from an idea to a native Swift app prototype in a few days.Second, security.
We’ve long led the industry in security. Now, our Security Operations Center agents automatically triage tens of thousands of unstructured threat reports each month, reducing threat mitigation time by more than 90%. And we have built and actively use Gemini-based AI agents (like CodeMender) to find and, importantly, fix critical software flaws.Third, our operations.
For the launch of Gemini in Chrome, our marketing teams used our models to rapidly generate thousands of variations of our creative assets, which would historically take weeks. Using AI led to 70% faster turnaround and a 20% increase in conversions, getting us to market faster and more effectively.Congratulations to our Google Cloud team, and a huge thanks to our partners who are building the future with us. We’ll have a lot more to share on how we’re bringing the latest technology to everyone at Google I/O on May 19.
Done. Just one step more.
Check your inbox to confirm your subscription.
You are already subscribed to our newsletter.
You can also subscribe with a
.png)
14 hours ago
English (United States) ·
French (France) ·