Home News Google Stealthily Surpasses in Enterprise AI: From 'Catch Up' to 'Catch Us'

Google Stealthily Surpasses in Enterprise AI: From 'Catch Up' to 'Catch Us'

April 20, 2025
EricMartin
0

Just a year ago, the buzz around Google and enterprise AI seemed stuck in neutral. Despite pioneering technologies like the Transformer, the tech giant appeared to be lagging behind, eclipsed by the viral success of OpenAI, the coding prowess of Anthropic, and Microsoft's aggressive push into the enterprise market.

But fast forward to last week at Google Cloud Next 2025 in Las Vegas, and the scene was markedly different. A confident Google, armed with top-performing models, robust infrastructure, and a clear enterprise strategy, announced a dramatic turnaround. In a closed-door analyst meeting with senior Google executives, one analyst encapsulated the mood: "This feels like the moment when Google went from 'catch up' to 'catch us.'"

Google's Leap Forward

This sentiment—that Google has not only caught up but surged ahead of OpenAI and Microsoft in the enterprise AI race—was palpable throughout the event. And it's not just marketing hype. Over the past year, Google has intensely focused on execution, turning its technological prowess into a high-performance, integrated platform that's quickly winning over enterprise decision-makers. From the world's most powerful AI models running on highly efficient custom silicon to a burgeoning ecosystem of AI agents tailored for real-world business challenges, Google is making a strong case that it was never truly lost but rather undergoing a period of deep, foundational development.

With its integrated stack now operating at full capacity, Google appears poised to lead the next phase of the enterprise AI revolution. In my interviews with Google executives at Next, they emphasized Google's unique advantages in infrastructure and model integration—advantages that competitors like OpenAI, Microsoft, or AWS would find challenging to replicate.

The Shadow of Doubt: Acknowledging the Recent Past

To fully appreciate Google's current momentum, it's essential to acknowledge the recent past. Google invented the Transformer architecture, sparking the modern revolution in large language models (LLMs), and began investing in specialized AI hardware (TPUs) a decade ago, which now drives industry-leading efficiency. Yet, inexplicably, just two and a half years ago, Google found itself playing defense.

OpenAI's ChatGPT captured the public's imagination and enterprise interest at a breathtaking pace, becoming the fastest-growing app in history. Competitors like Anthropic carved out niches in areas like coding. Meanwhile, Google's public moves often seemed tentative or flawed. The infamous Bard demo fumbles in 2023 and the controversy over its image generator producing historically inaccurate depictions fed a narrative of a company potentially hindered by internal bureaucracy or overcorrection on alignment. Google seemed lost, echoing its initial slowness in the cloud competition where it remained a distant third in market share behind Amazon and Microsoft.

The Pivot: A Conscious Decision to Lead

Behind the scenes, however, a significant shift was occurring, driven by a deliberate decision at the highest levels to reclaim leadership. Mat Velloso, VP of product for Google DeepMind’s AI Developer Platform, sensed this pivotal moment upon joining Google in February 2024, after leaving Microsoft. "When I came to Google, I spoke with Sundar [Pichai], I spoke with several leaders here, and I felt like that was the moment where they were deciding, okay, this [generative AI] is a thing the industry clearly cares about. Let’s make it happen," Velloso shared during an interview at Next last week.

This renewed push wasn't hampered by the feared "brain drain" that some outsiders believed was depleting Google. Instead, the company doubled down on execution in early 2024, marked by aggressive hiring, internal unification, and customer traction. While competitors made splashy hires, Google retained its core AI leadership, including DeepMind CEO Demis Hassabis and Google Cloud CEO Thomas Kurian, providing stability and deep expertise.

Moreover, talent began flowing towards Google’s focused mission. Logan Kilpatrick, for instance, returned to Google from OpenAI, drawn by the opportunity to build foundational AI within the company. He joined Velloso in what he described as a "zero to one experience," tasked with building developer traction for Gemini from the ground up. "It was like the team was me on day one... we actually have no users on this platform, we have no revenue. No one is interested in Gemini at this moment," Kilpatrick recalled of the starting point. Leaders like Josh Woodward, who helped start AI Studio and now leads the Gemini App and Labs, and Noam Shazeer, a key co-author of the original "Attention Is All You Need" Transformer paper, also returned to the company in late 2024 as a technical co-lead for the crucial Gemini project.

Pillar 1: Gemini 2.5 and the Era of Thinking Models

While the enterprise mantra has shifted to "it's not just about the model," having the best-performing LLM remains a significant achievement and a powerful validator of a company's superior research and efficient technology architecture. With the release of Gemini 2.5 Pro just weeks before Next '25, Google decisively claimed this mantle. It quickly topped the independent Chatbot Arena leaderboard, significantly outperforming even OpenAI’s latest GPT-4o variant, and aced notoriously difficult reasoning benchmarks like Humanity’s Last Exam. As Pichai stated in the keynote, "It’s our most intelligent AI model ever. And it is the best model in the world." The model had driven an 80 percent increase in Gemini usage within a month, he Tweeted separately.

For the first time, demand for Gemini was on fire. What impressed me, aside from Gemini 2.5 Pro’s raw intelligence, was its demonstrable reasoning. Google has engineered a "thinking" capability, allowing the model to perform multi-step reasoning, planning, and even self-reflection before finalizing a response. The structured, coherent chain-of-thought (CoT) – using numbered steps and sub-bullets – avoids the rambling or opaque nature of outputs from other models from DeepSeek or OpenAI. For technical teams evaluating outputs for critical tasks, this transparency allows validation, correction, and redirection with unprecedented confidence.

But more importantly for enterprise users, Gemini 2.5 Pro also dramatically closed the gap in coding, which is one of the biggest application areas for generative AI. In an interview with VentureBeat, CTO Fiona Tan of leading retailer Wayfair said that after initial tests, the company found it "stepped up quite a bit" and was now "pretty comparable" to Anthropic’s Claude 3.7 Sonnet, previously the preferred choice for many developers.

Google also added a massive 1 million token context window to the model, enabling reasoning across entire codebases or lengthy documentation, far exceeding the capabilities of the models of OpenAI or Anthropic. (OpenAI responded this week with models featuring similarly large context windows, though benchmarks suggest Gemini 2.5 Pro retains an edge in overall reasoning). This advantage allows for complex, multi-file software engineering tasks.

Complementing Pro is Gemini 2.5 Flash, announced at Next '25 and released just yesterday. Also a "thinking" model, Flash is optimized for low latency and cost-efficiency. You can control how much the model reasons and balance performance with your budget. This tiered approach further reflects the "intelligence per dollar" strategy championed by Google executives.

Velloso showed a chart revealing that across the intelligence spectrum, Google models offer the best value. "If we had this conversation one year ago… I would have nothing to show," Velloso admitted, highlighting the rapid turnaround. "And now, like, across the board, we are, if you’re looking for whatever model, whatever size, like, if you’re not Google, you’re losing money." Similar charts have been updated to account for OpenAI’s latest model releases this week, all showing the same thing: Google’s models offer the best intelligence per dollar. See below:

For any given price, Google’s models offer more intelligence than other models, about 90 percent of the time. Source: Pierre Bongrand.

Wayfair’s Tan also observed promising latency improvements with 2.5 Pro: "Gemini 2.5 came back faster," making it viable for "more customer-facing sort of capabilities," she said, something she said hasn’t been the case before with other models. Gemini could become the first model Wayfair uses for these customer interactions, she said.

The Gemini family’s capabilities extend to multimodality, integrating seamlessly with Google’s other leading models like Imagen 3 (image generation), Veo 2 (video generation), Chirp 3 (audio), and the newly announced Lyria (text-to-music), all accessible via Google’s platform for Enterprise users, Vertex. Google is the only company that offers its own generative media models across all modalities on its platform. Microsoft, AWS, and OpenAI have to partner with other companies to do this.

Pillar 2: Infrastructure Prowess – The Engine Under the Hood

The ability to rapidly iterate and efficiently serve these powerful models stems from Google’s arguably unparalleled infrastructure, honed over decades of running planet-scale services. Central to this is the Tensor Processing Unit (TPU).

At Next '25, Google unveiled Ironwood, its seventh-generation TPU, explicitly designed for the demands of inference and "thinking models." The scale is immense, tailored for demanding AI workloads: Ironwood pods pack over 9,000 liquid-cooled chips, delivering a claimed 42.5 exaflops of compute power. Google’s VP of ML Systems Amin Vahdat said on stage at Next that this is "more than 24 times" the compute power of the world’s current #1 supercomputer.

Google stated that Ironwood offers 2x perf/watt relative to Trillium, the previous generation of TPU. This is significant since enterprise customers increasingly say energy costs and availability constrain large-scale AI deployments.

Google Cloud CTO Will Grannis emphasized the consistency of this progress. Year over year, Google is making 10x, 8x, 9x, 10x improvements in its processors, he told VentureBeat in an interview, creating what he called a "hyper Moore’s law" for AI accelerators. He said customers are buying Google’s roadmap, not just its technology.

Google’s position fueled this sustained TPU investment. It needs to efficiently power massive services like Search, YouTube, and Gmail for more than 2 billion users. This necessitated developing custom, optimized hardware long before the current generative AI boom. While Meta operates at a similar consumer scale, other competitors lacked this specific internal driver for decade-long, vertically integrated AI hardware development.

Now these TPU investments are paying off because they are driving the efficiency not only for its own apps but also allowing Google to offer Gemini to other users at a better intelligence per dollar, all things being equal.

Why can’t Google’s competitors buy efficient processors from Nvidia, you ask? It’s true that Nvidia’s GPU processors dominate the process pre-training of LLMs. But market demand has pushed up the price of these GPUs, and Nvidia takes a healthy cut for itself as profit. This passes significant costs along to users of its chips. And also, while pre-training has dominated the usage of AI chips so far, this is changing now that enterprises are actually deploying these applications. This is where "inference" comes in, and here TPUs are considered more efficient than GPUs for workloads at scale.

When you ask Google executives where their main technology advantage in AI comes from, they usually fall back to the TPU as the most important. Mark Lohmeyer, the VP who runs Google’s computing infrastructure, was unequivocal: TPUs are "certainly a highly differentiated part of what we do… OpenAI, they don’t have those capabilities."

Significantly, Google presents TPUs not in isolation, but as part of the wider, more complex enterprise AI architecture. For technical insiders, it’s understood that top-tier performance hinges on integrating increasingly specialized technology breakthroughs. Many updates were detailed at Next. Vahdat described this as a "supercomputing system," integrating hardware (TPUs, the latest Nvidia GPUs like Blackwell and upcoming Vera Rubin, advanced storage like Hyperdisk Exapools, Anywhere Cache, and Rapid Storage) with a unified software stack. This software includes Cluster Director for managing accelerators, Pathways (Gemini’s distributed runtime, now available to customers), and bringing optimizations like vLLM to TPUs, allowing easier workload migration for those previously on Nvidia/PyTorch stacks. This integrated system, Vahdat argued, is why Gemini 2.0 Flash achieves 24 times higher intelligence per dollar, compared to GPT-4o.

Google is also extending its physical infrastructure reach. Cloud WAN makes Google’s low-latency 2-million-mile private fiber network available to enterprises, promising up to 40% faster performance and 40% lower total cost of ownership (TCO) compared to customer-managed networks.

Furthermore, Google Distributed Cloud (GDC) allows Gemini and Nvidia hardware (via a Dell partnership) to run in sovereign, on-premises, or even air-gapped environments – a capability Nvidia CEO Jensen Huang lauded as "utterly gigantic" for bringing state-of-the-art AI to regulated industries and nations. At Next, Huang called Google’s infrastructure the best in the world: "No company is better at every single layer of computing than Google and Google Cloud," he said.

Pillar 3: The Integrated Full Stack – Connecting the Dots

Google’s strategic advantage grows when considering how these models and infrastructure components are woven into a cohesive platform. Unlike competitors, which often rely on partnerships to bridge gaps, Google controls nearly every layer, enabling tighter integration and faster innovation cycles.

So why does this integration matter, if a competitor like Microsoft can simply partner with OpenAI to match infrastructure breadth with LLM model prowess? The Googlers I talked with said it makes a huge difference, and they came up with anecdotes to back it up.

Take the significant improvement of Google’s enterprise database BigQuery. The database now offers a knowledge graph that allows LLMs to search over data much more efficiently, and it now boasts more than five times the customers of competitors like Snowflake and Databricks. Yasmeen Ahmad, Head of Product for Data Analytics at Google Cloud, said the vast improvements were only possible because Google’s data teams were working closely with the DeepMind team. They worked through use cases that were hard to solve, and this led to the database providing 50 percent more accuracy based on common queries, at least according to Google’s internal testing, in getting to the right data than the closest competitors, Ahmad told VentureBeat in an interview. Ahmad said this sort of deep integration across the stack is how Google has "leapfrogged" the industry.

This internal cohesion contrasts sharply with the "frenemies" dynamic at Microsoft. While Microsoft partners with OpenAI to distribute its models on the Azure cloud, Microsoft is also building its own models. Mat Velloso, the Google executive who now leads the AI developer program, left Microsoft after getting frustrated trying to align Windows Copilot plans with OpenAI’s model offerings. "How do you share your product plans with another company that’s actually competing with you… The whole thing is a contradiction," he recalled. "Here I sit side by side with the people who are building the models."

This integration speaks to what Google leaders see as their core advantage: its unique ability to connect deep expertise across the full spectrum, from foundational research and model building to "planet-scale" application deployment and infrastructure design.

Vertex AI serves as the central nervous system for Google’s enterprise AI efforts. And the integration goes beyond just Google’s own offerings. Vertex’s Model Garden offers over 200 curated models, including Google’s, Meta’s Llama 4, and numerous open-source options. Vertex provides tools for tuning, evaluation (including AI-powered Evals, which Grannis highlighted as a key accelerator), deployment, and monitoring. Its grounding capabilities leverage internal AI-ready databases alongside compatibility with external vector databases. Add to that Google’s new offerings to ground models with Google Search, the world’s best search engine.

Integration extends to Google Workspace. New features announced at Next '25, like "Help Me Analyze" in Sheets (yes, Sheets now has an "=AI" formula), Audio Overviews in Docs and Workspace Flows, further embed Gemini’s capabilities into daily workflows, creating a powerful feedback loop for Google to use to improve the experience.

While driving its integrated stack, Google also champions openness where it serves the ecosystem. Having driven Kubernetes adoption, it’s now promoting JAX for AI frameworks and now open protocols for agent communication (A2A) alongside support for existing standards (MCP). Google is also offering hundreds of connectors to external platforms from within Agentspace, which is Google’s new unified interface for employees to find and use agents. This hub concept is compelling. The keynote demonstration of Agentspace (starting at 51:40) illustrates this. Google offers users pre-built agents, or employees or developers can build their own using no-code AI capabilities. Or they can pull in agents from the outside via A2A connectors. It integrates into the Chrome browser for seamless access.

Pillar 4: Focus on Enterprise Value and the Agent Ecosystem

Perhaps the most significant shift is Google’s sharpened focus on solving concrete enterprise problems, particularly through the lens of AI agents. Thomas Kurian, Google Cloud CEO, outlined three reasons customers choose Google: the AI-optimized platform, the open multi-cloud approach allowing connection to existing IT, and the enterprise-ready focus on security, sovereignty, and compliance.

Agents are key to this strategy. Aside from AgentSpace, this also includes:

  • Building Blocks: The open-source Agent Development Kit (ADK), announced at Next, has already seen significant interest from developers. The ADK simplifies creating multi-agent systems, while the proposed Agent2Agent (A2A) protocol aims to ensure interoperability, allowing agents built with different tools (Gemini ADK, LangGraph, CrewAI, etc.) to collaborate. Google’s Grannis said that A2A anticipates the scale and security challenges of a future with potentially hundreds of thousands of interacting agents.
  • Purpose-built Agents: Google showcased expert agents integrated into Agentspace (like NotebookLM, Idea Generation, Deep Research) and highlighted five key categories gaining traction: Customer Agents (powering tools like Reddit Answers, Verizon’s support assistant, Wendy’s drive-thru), Creative Agents (used by WPP, Brandtech, Sphere), Data Agents (driving insights at Mattel, Spotify, Bayer), Coding Agents (Gemini Code Assist), and Security Agents (integrated into the new Google Unified Security platform).

This comprehensive agent strategy appears to be resonating. Conversations with executives at three other large enterprises this past week, also speaking anonymously due to competitive sensitivities, echoed this enthusiasm for Google’s agent strategy. Google Cloud COO Francis DeSouza confirmed in an interview: "Every conversation includes AI. Specifically, every conversation includes agents."

Kevin Laughridge, an executive at Deloitte, a big user of Google’s AI products, and a distributor of them to other companies, described the agent market as a "land grab" where Google’s early moves with protocols and its integrated platform offer significant advantages. "Whoever is getting out first and getting the most agents that actually deliver value – is who is going to win in this race," Laughridge said in an interview. He said Google’s progress was "astonishing," noting that custom agents Deloitte built just a year ago could now be replicated "out of the box" using Agentspace. Deloitte itself is building 100 agents on the platform, targeting mid-office functions like finance, risk, and engineering, he said.

The customer proof points are mounting. At Next, Google cited "500 plus customers in production" with generative AI, up from just "dozens of prototypes" a year ago. If Microsoft was perceived as way ahead a year ago, that doesn’t seem so obviously the case anymore. Given the PR war from all sides, it’s difficult to say who is really winning right now definitively. Metrics vary. Google’s 500 number isn’t directly comparable to the 400 case studies Microsoft promotes (and Microsoft, in response, told VentureBeat at press time that it plans to update this public count to 600 shortly, underscoring the intense marketing). And if Google’s distribution of AI through its apps is significant, Microsoft’s Copilot distribution through its 365 offering is equally impressive. Both are now hitting millions of developers through APIs.

But examples abound of Google’s traction:

  • Wendy’s: Deployed an AI drive-thru system to thousands of locations in just one year, improving employee experience and order accuracy. Google Cloud CTO Will Grannis noted that the AI system is capable of understanding slang and filtering out background noise, significantly reducing the stress of live customer interactions. That frees up staff to focus on food prep and quality — a shift Grannis called "a great example of AI streamlining real-world operations."
  • Salesforce: Announced a major expansion, enabling its platform to run on Google Cloud for the first time (beyond AWS), citing Google’s ability to help them "innovate and optimize."
  • Honeywell & Intuit: Companies previously strongly associated with Microsoft and AWS, respectively, now partnering with Google Cloud on AI initiatives.
  • Major Banks (Deutsche Bank, Wells Fargo): Leveraging agents and Gemini for research, analysis, and modernizing customer service.
  • Retailers (Walmart, Mercado Libre, Lowe’s): Using search, agents, and data platforms.

This enterprise traction fuels Google Cloud’s overall growth, which has outpaced AWS and Azure for the last three quarters. Google Cloud reached a $44 billion annualized run rate in 2024, up from just $5 billion in 2018.

Navigating the Competitive Waters

Google’s ascent doesn’t mean competitors are standing still. OpenAI’s rapid releases this week of GPT-4.1 (focused on coding and long context) and the o-series (multimodal reasoning, tool use) demonstrate OpenAI’s continued innovation. Moreover, OpenAI’s new image generation feature update in GPT-4o fueled massive growth over just the last month, helping ChatGPT reach 800 million users. Microsoft continues to leverage its vast enterprise footprint and OpenAI partnership, while Anthropic remains a strong contender, particularly in coding and safety-conscious applications.

However, it’s indisputable that Google’s narrative has improved remarkably. Just a year ago, Google was viewed as a stodgy, halting, blundering competitor that perhaps was about to blow its chance at leading AI at all. Instead, its unique, integrated stack and corporate steadfastness have revealed something else: Google possesses world-class capabilities across the entire spectrum – from chip design (TPUs) and global infrastructure to foundational model research (DeepMind), application development (Workspace, Search, YouTube), and enterprise cloud services (Vertex AI, BigQuery, Agentspace). "We’re the only hyperscaler that’s in the foundational model conversation," deSouza stated flatly. This end-to-end ownership allows for optimizations (like "intelligence per dollar") and integration depth that partnership-reliant models struggle to match. Competitors often need to stitch together disparate pieces, potentially creating friction or limiting innovation speed.

Google’s Moment is Now

While the AI race remains dynamic, Google has assembled all these pieces at the precise moment the market demands them. As Deloitte’s Laughridge put it, Google hit a point where its capabilities aligned perfectly "where the market demanded it." If you were waiting for Google to prove itself in enterprise AI, you may have missed the moment — it already has. The company that invented many of the core technologies powering this revolution appears to have finally caught up – and more than that, it is now setting the pace that competitors need to match.

In the video below, recorded right after Next, AI expert Sam Witteveen and I break down the current landscape and emerging trends, and why Google’s AI ecosystem feels so strong:

Related article
MCP Standardizes AI Connectivity with Tools and Data: A New Protocol Emerges MCP Standardizes AI Connectivity with Tools and Data: A New Protocol Emerges If you're diving into the world of artificial intelligence (AI), you've probably noticed how crucial it is to get different AI models, data sources, and tools to play nicely together. That's where the Model Context Protocol (MCP) comes in, acting as a game-changer in standardizing AI connectivity. T
Explore Earth Virtually: ChatGPT and Google Earth Vacation Planner Explore Earth Virtually: ChatGPT and Google Earth Vacation Planner Ever felt the urge to escape the daily grind but found yourself stumped on where to go? Let's dive into a cool way to plan your next getaway without even stepping outside your door. By harnessing the power of ChatGPT and Google Earth, you can embark on a virtual vacation that's both exciting and rel
ChatGPT's Unsolicited Use of User Names Sparks 'Creepy' Concerns Among Some ChatGPT's Unsolicited Use of User Names Sparks 'Creepy' Concerns Among Some Some users of ChatGPT have recently encountered an odd new feature: the chatbot occasionally uses their name while working through problems. This wasn't part of its usual behavior before, and many users report that ChatGPT mentions their names without ever being told what to call them. Opinions on
Comments (10)
0/200
JeffreyHarris April 20, 2025 at 10:35:09 AM GMT

Google's sneaky move into enterprise AI is pretty cool. From playing catch-up to leading the pack, they've really turned things around. The tech is solid but the interface could be more user-friendly. Still, it's impressive how fast they've moved! 🚀

WillieJones April 22, 2025 at 12:13:31 AM GMT

グーグルのエンタープライズAIへの静かな進出はかなりクールですね。追いつくことからリードするまで、本当に状況を変えました。技術はしっかりしていますが、インターフェースがもう少しユーザーフレンドリーだといいですね。それでも、進むスピードが印象的です!🚀

WillBaker April 22, 2025 at 5:12:49 AM GMT

구글이 기업용 AI로 몰래 진출한 것이 꽤 멋지네요. 따라잡기에서 선두로 나아가 정말 상황을 바꿨어요. 기술은 훌륭하지만 인터페이스가 좀 더 사용자 친화적이면 좋겠어요. 그래도 이렇게 빠르게 움직인 것이 인상적이에요! 🚀

CharlesMartinez April 22, 2025 at 6:10:16 AM GMT

A jogada furtiva do Google no AI empresarial é bem legal. De tentar alcançar a liderar o grupo, eles realmente mudaram as coisas. A tecnologia é sólida, mas a interface poderia ser mais amigável ao usuário. Ainda assim, é impressionante como eles se moveram rápido! 🚀

JasonRoberts April 22, 2025 at 7:22:36 AM GMT

El movimiento sigiloso de Google en la IA empresarial es bastante genial. De ir a la zaga a liderar el grupo, realmente han dado la vuelta a la situación. La tecnología es sólida pero la interfaz podría ser más amigable para el usuario. Aún así, es impresionante lo rápido que se han movido! 🚀

FredAnderson April 22, 2025 at 3:53:12 PM GMT

Google's move into enterprise AI is impressive. From playing catch-up to leading the pack, they've really stepped up their game. But the interface could use some work; it's a bit clunky. Still, I'm excited to see where they go from here! 🚀

OR