
Google Cloud Next 2026 Focus: AI Agents Enter Scale; Inference Chips Become Independent Growth Curve
Google Cloud Next 2026 sent a key signal: the enterprise AI battlefield has shifted from 'how to experiment' to 'how to govern and scale deployment'. 75% of cloud customers are using AI products, and Gemini paid monthly active users grew 40% quarter-over-quarter in a single quarter. TPU was split for the first time into two product lines for training and inference, while platform restructuring targets an 'agentic enterprise operating system'. JPMorgan, Bank of America, and Citigroup all maintained Buy ratings after the event, with the highest target price set at $405
Google's annual cloud computing conference, Cloud Next 2026, sent a clear signal: the enterprise AI battlefield has shifted from "how to experiment" to "how to govern and scale deployment," with Google's response being a complete vertical stack ranging from chips to platforms. This conference was not merely a product launch showcase but marked a critical inflection point where agentic AI is moving beyond proof-of-concept toward enterprise-grade production deployment.
According to Trading Wind Desk, JPMorgan analyst Doug Anmuth wrote after the event: "This shift from experimentation to deployment may be the strongest evidence that agent AI is bridging the gap between proof-of-concept and enterprise workloads." Data from the demand side confirms this judgment: Google's first-party models now handle 16 billion tokens per minute via direct API connections, a significant increase from 10 billion last quarter; approximately 75% of Cloud customers are using its AI products; and Gemini Enterprise paid monthly active users grew 40% quarter-over-quarter in the first quarter.
Three institutions—JPMorgan, Bank of America Securities, and Citigroup Research—all maintained Buy ratings on Alphabet following the event, with target prices of $395, $370, and $405 respectively. The shared logic is that Cloud revenue growth continues to outpace advertising, and the combination of "Gemini models + self-developed TPU + enterprise orchestration platform" is building a differentiated moat and is expected to become a more direct driver of stock price. Meanwhile, Sundar Pichai outlined a capital expenditure range of $175 billion to $185 billion for 2026 in his keynote address, and the market remains highly attentive to the capex trajectory around the earnings window.
Enterprise Customer Questions Have Shifted: From "How to Try" to "How to Manage"
If Cloud Next in the past two years served as a showcase for technical capabilities, this year's theme has switched to how to promote AI from experimental deployments by a few pioneering enterprises to scalable, governable, and cost-controlled production workloads.
JPMorgan traced this evolution path in its research report: 2024 focused on integrating Gemini with Workspace and early agent exploration; 2025 emphasized A2A protocols and the seventh-generation TPU Ironwood; and by 2026, several key themes surrounding Agentic Cloud, data usability, AI infrastructure cost efficiency, and security converged on one outcome—moving agents from pilot projects to sustainable production deployment.
Citigroup Research analyst Ronald Josey stated more directly: As managers begin "managing multiple agents across workflows," enterprises are moving from "knowing how to use models" to "using agents to transform processes." Google Cloud is betting on this migration direction, positioning it as "the key operating system for the agentic enterprise."
This context also explains why the conference information density concentrated on two levels: compute and network forms for agent workflows, and upgrading the platform to an "agent factory." Google chose not to release any financial updates at the event, but instead used customer usage metrics to prove products are running authentically in production environments—including that about 75% of new code within Google itself is now AI-generated and reviewed by engineers, and threat resolution times on the security side have been shortened by over 90%.
TPU Gen 8: Inference Split from Training to Become Independent Capital Narrative
The most structurally significant change on the hardware side at this conference was the first-time split of the eighth-generation TPU into two independent product lines: TPU 8t for high-throughput training workloads, and TPU 8i positioned as a dedicated chip "optimized from the ground up for real-time inference."
The logical chain behind this "forked architecture" was articulated most clearly in the JPMorgan research report: TPU 8t expands clusters to over one million chips per cluster via the new Virgo Network fabric, achieving peak performance roughly three times that of the previous generation Ironwood, with the goal of compressing training time for trillion-parameter frontier models. Meanwhile, TPU 8i adopts a new boardfly network topology, increasing on-chip SRAM by about threefold, with the core objective of breaking through latency and memory bottlenecks encountered during scaled agentic inference. Citigroup's report added an efficiency dimension: TPU 8i reduces latency by approximately five times compared to TPU 7 and improves performance-per-dollar by about 80%.
JPMorgan's inference logic is worth noting: since inference no longer "reuses training chips" but requires specialized ASIC optimization, it indicates Google judges that inference compute demand has grown large enough to justify separate silicon development and separate capital allocation. Revenue opportunities have thus undergone a structural shift—not just following training, but increasingly coming from sustained consumption on the inference side, forming an independent growth curve.
Notably, all three reports mentioned that management did not discuss the possibility of selling TPU externally at the conference, meaning this hardware route currently serves more the logic of "self-use plus selling cloud services" rather than evolving into an independent hardware commercialization narrative.
Platform Restructuring: Vertex AI "Upgraded" to Unified Governance Entry for Enterprise Agents
Beyond hardware, platform restructuring represented another structural change worthy of attention at this conference. Google launched the Gemini Enterprise Agent Platform, which JPMorgan described as effectively "superseding Vertex AI"—consolidating enterprise build, orchestration, governance, and security into a unified entry point rather than scattered functional modules.
Bank of America Securities broke down this restructuring into three layers. At the infrastructure layer, AI Hypercomputer integrates GPU/TPU, high-speed networks, storage, and optimized software into a single architecture covering the full lifecycle from training to inference. At the platform layer, capabilities are organized around four dimensions: "build/scale/govern/optimize," including low-code/no-code agent creation, centralized management, cross-ecosystem orchestration (linking Google Workspace, Microsoft 365, and third-party applications), and built-in observability and traceability. At the application layer, Workspace Intelligence pushes agent capabilities down to high-frequency office entry points like Gmail, Docs, and Chat, allowing multi-step tasks to be executed across applications.

Citigroup's interpretation angle differed slightly, emphasizing that the platform's key value lies in "enabling enterprises to run processes with multiple agents within the same management system." From a product philosophy perspective, this means: the threshold for deploying agents at scale no longer depends solely on a company's technical depth, but rather on whether the platform's pre-built capabilities are standardized enough to allow more enterprises to bypass custom engineering and move directly to production deployment.
Google Backs Narrative with Internal Data: "Full-Stack AI" Proven in Production
The conference did not disclose financial data; instead, Google chose to support the narrative that "agents have entered production" with internally quantifiable cases. Citigroup grouped these cases into four dimensions:
On the R&D side, about 75% of new code is generated by AI and approved by engineers; Citigroup provided a longitudinal comparison—this ratio was about 50% in October 2025 and about 30% in Q1 2025, showing significant penetration speed. One code migration project was described as being completed six times faster than a year ago.
On the marketing and content production side, turnaround time from concept to video assets accelerated by about 70%, accompanied by a conversion rate increase of about 20%.
On the security side, Google Cloud automatically processes tens of thousands of unstructured threat reports monthly, shortening threat mitigation time by over 90%; security capabilities rely on integrated Wiz and Mandiant solutions to form a differentiated product suite. Citigroup also noted that AI has compressed "average time-to-exploit" to "negative seven days," meaning patches often haven't even been released before attacks occur, further amplifying the strategic value of automated security orchestration.
On the customer service side, YouTube deployed an AI voice agent within six weeks, covering call scenarios for NFL Sunday Ticket and YouTube TV; Citigroup emphasized its low latency, accuracy, and bilingual capabilities.
The common function of these cases across the three reports is to distinguish "real enterprise-side load" from "showcase demos," supporting the judgment that there is upside potential for Cloud's quarterly performance.
$175B–$185B Capex Range: Not "Hold Steady," But Not "Peak Yet"
Sundar Pichai's announcement of a 2026 capital expenditure range of $175 billion to $185 billion in his keynote was the only statement regarding financial magnitude at the conference and also the topic with relatively greater divergence among the three reports.
JPMorgan's interpretation leaned pragmatic: publicly mentioning this range increases the probability of maintaining existing guidance unchanged in next week's earnings report, rather than confirming a capex ceiling. Its own forecast is approximately $181 billion for 2026 and $226 billion for 2027 (about 25% year-over-year growth), roughly 12% higher than market consensus. The report also placed another counter-trend line on the table: both Amin Vahdat and Jeff Dean emphasized at the conference that AI remains supply-constrained, implying that the capex trajectory "may still have room for upward movement," and the conclusion that "the range represents the ceiling" does not hold.
Bank of America Securities listed Capex/FCF pressure directly in its downside risk list: AI investment driving up capital expenditures and depressing free cash flow is one of the most direct factors pressuring profit margins.
The consensus among the three reports is: Cloud Next addressed whether Google has agentic AI products and infrastructure; the question for the coming quarters is whether these investments can deliver Cloud's growth and margin expectations without significantly sacrificing cash flow.
Three Investment Banks Maintain Buy Ratings, Though Risk Lists Differ
In terms of investment conclusions, all three reports maintained Buy ratings, though valuation anchors and emphasis on arguments varied.
JPMorgan maintained Overweight with a 12-month target price of $395, based on approximately 29 times its 2027 GAAP EPS forecast of $13.51; the report lists Alphabet as a "top overall pick," supported not only by bets on cloud but also by runway remaining in Search and YouTube advertising, continued expansion in non-advertising businesses, and option value provided by Waymo.
Bank of America Securities maintained Buy with a target price of $370, based on 27 times 2027 core GAAP EPS plus cash per share; the report continuously increased Cloud's weight in SOTP analysis and provided a reference metric estimating a market cap contribution of about $1.2 trillion based on 10 times revenue, arguing that cloud margin expansion and AI asset monetization space support higher multiples.
Citigroup Research maintained Buy with the highest target price of $405, corresponding to approximately 29 times 2027 GAAP EPS of $13.92; the report attributes the premium to two points: accelerated revenue growth driven by TPU and Gemini demand for Google Cloud, and search business resilience driven by strong query volume.
Regarding risks, all three reports mentioned potential pressures from intensifying AI competition and search traffic diversion; JPMorgan and Bank of America Securities separately listed EU DMA compliance pressure. Bank of America Securities cited "slower-than-expected integration of LLMs in search potentially negatively impacting search revenue" as the primary short-term uncertainty, with the next validation node returning to Q1 earnings disclosed after market close on April 29.
