What Google Cloud announced in AI this month – and how it helps you (original) (raw)

Editor’s note: Want to keep up with the latest from Google Cloud? Check back here for a monthly recap of our latest updates, announcements, resources, events, learning opportunities, and more.


This month, we're buzzing with the news of AP2, the Agent Payments Protocol, an open protocol developed with 60+ leading payments and technology companies. AP2 is a blueprint for interoperable, AI-native commerce, designed to work seamlessly as an extension of the Agent2Agent (A2A) protocol and Model Context Protocol (MCP).

We’ve also seen “bananas” momentum with our generative media models, especially through nano-banana, or Gemini 2.5 Flash Image. Let’s dive in!

Top hits:

https://storage.googleapis.com/gweb-cloudblog-publish/images/1_wxU7fcB.max-2000x2000.png

https://storage.googleapis.com/gweb-cloudblog-publish/images/2_gxUYwEG.max-1500x1500.png

News you can use:

https://storage.googleapis.com/gweb-cloudblog-publish/images/3_YVRxV4G.max-1800x1800.png

      1. /security:analyze performs a comprehensive scan right in your local repository, with support for GitHub pull requests coming soon. This makes security a natural part of your development cycle.
        1. /deploy deploys your application to Cloud Run, our fully managed serverless platform, in just a few minutes.

Stay tuned for monthly updates on Google Cloud’s AI announcements, news, and best practices. For a deeper dive into the latest from Google Cloud, read our weekly updates, The Overwhelmed Person’s Guide to Google Cloud.

https://storage.googleapis.com/gweb-cloudblog-publish/images/maxresdefault_PDdX617.max-1300x1300.jpg

https://storage.googleapis.com/gweb-cloudblog-publish/images/maxresdefault_zX15ZOm.max-1300x1300.jpg

News you can use

https://storage.googleapis.com/gweb-cloudblog-publish/images/2_0eLq7ZQ.max-2000x2000.png


July

Do you know since its preview launch on Vertex AI in June, enterprise customers have already generated over 6 million videos?

This July, we focused on helping you build what's next, and do it with confidence. We announced Veo 3 and Veo 3 Fast on Vertex AI to bring your ideas to life. This is exciting, especially for marketers, who want to rapidly iterate on high-quality creative. Plus, we’ve pulled together over 25+ of our best guides – from model deployment to building gen AI apps – to help you find ways to get started.

Top hits

https://storage.googleapis.com/gweb-cloudblog-publish/images/maxresdefault-8.max-1300x1300.jpg

News you can use

Stay tuned for monthly updates on Google Cloud’s AI announcements, news, and best practices. For a deeper dive into the latest from Google Cloud, read our weekly updates, The Overwhelmed Person’s Guide to Google Cloud.


June

From the command line to cinematic screens, June was about making AI practical and powerful. Developers can now access Gemini right from their terminal with our new command-line interface, Gemini CLI, and creators can produce stunning, narrative-driven video that combines audio and visuals with Veo 3. Plus, we published several helpful guides and resources for you to bookmark and use as you start building.

Top hits: Gemini comes to everyone – from your command line to Vertex AI

The story unfolds: You dream it, Veo creates it

Any creative class will tell you the same axiom of truth: A great story shows rather than tells. We gave our text-to-video model, Veo 3, a simple, evocative prompt to show how it can turn a creative spark into a cinematic scene. Veo 3 is now available for all Google Cloud customers and partners in public preview on Vertex AI.

The prompt: "A medium shot frames an old sailor, his knitted blue sailor hat casting a shadow over his eyes, a thick grey beard obscuring his chin. He holds his pipe in one hand, gesturing with it towards the churning, grey sea beyond the ship's railing. 'This ocean, it's a force, a wild, untamed might. And she commands your awe, with every breaking light'"

The result:

https://storage.googleapis.com/gweb-cloudblog-publish/images/maxresdefault_g415WGs.max-1300x1300.jpg

AI agents – they’re taking autonomous action, but we need to keep them secure

As AI moves from answering questions to taking action, securing these autonomous systems is paramount. This month, we're not only highlighting how to build powerful agents but also how to implement a security-first approach to their deployment.

Exec perspective: Earlier this month, Anton Chuvakin, security advisor for Google Cloud’s Office of the CISO, discussed a new Google report on securing AI agents, and the new security paradigm they demand. Here’s a snippet: With great power comes a great responsibility for agent developers. To help mitigate the potential risks posed by rogue agent actions, we should invest in a new field of study focused specifically on securing agent systems.

Keeping the conversation going: What are the unique security threats posed by agentic AI, and what can leaders do to start securing their workloads? To help answer that question, we sat down with Archana Ramamoorthy, Senior Director, Google Cloud Security, to ask her how businesses should protect their AI workloads by making security more collaborative. Here’s a snippet:

“Just as leaders are getting a handle on the security risks of generative AI, the ground is shifting again. The emergence of AI agents — which can execute tasks on your company's behalf — is demanding a new security paradigm: agent security. This new layer of autonomy can magnify blind spots in an organization's existing security posture…The bottom line is that we need to be prepared, and the whole organization should invest in keeping systems secure.”

News you can use

Build smarter, multimodal AI applications

Deploy your AI efficiently and at scale

https://storage.googleapis.com/gweb-cloudblog-publish/images/image2_cLQ1nrQ.max-1300x1300.png

Ground your AI in your enterprise data

Stay tuned

Stay tuned for monthly updates on Google Cloud’s AI announcements, news, and best practices. For a deeper dive into the latest from Google Cloud, read our weekly updates, The Overwhelmed Person’s Guide to Google Cloud.

https://storage.googleapis.com/gweb-cloudblog-publish/images/maxresdefault_sgK1KZC.max-1300x1300.jpg

Top announcements

Google I/O brought a fresh wave of tools from Google Cloud, all designed to help businesses and developers build what's next. These updates bring new ways for organizations to work with AI, code more easily, create media, and manage intelligent agents. Here are the highlights:

  1. We introduced new generative AI models for media, including Veo 3 for video, Imagen 4 for images, and Lyria 2 for music on Vertex AI. These models give you excellent ways to create visual and audio content from text prompts. Learn more in our blog here.
  2. We've expanded Gemini 2.5 Flash and Pro model capabilities to help enterprises build more sophisticated and secure AI-driven applications and agents. With thought summaries, businesses get clarity and auditability of a model’s raw thoughts — including key details and tool usage. The new Deep Think mode uses research techniques that enable the model to consider multiple hypotheses before responding.
  3. Gemini 2.5 is now powering all Gemini Code Assist editions! We also launched Jules, a new autonomous AI coding agent, now in public beta, designed to understand user intent and perform coding tasks like writing tests and fixing bugs
  4. Firebase Studio is a cloud-based, AI workspace powered by Gemini 2.5 that lets you turn your ideas into a full-stack app in minutes. Now you can import Figma designs directly into Firebase Studio using the builder.io plugin, and then add features and functionality using Gemini in Firebase without having to write any code.
  5. We're making AI application deployment significantly easier with Cloud Run, launching three key updates: first, you can now deploy applications built in Google AI Studio directly to Cloud Run with just a single button click; second, we enabled direct deployment of Gemma 3 models from AI Studio to Cloud Run, complete with GPU support for scalable, pay-per-use endpoints; and third, we've introduced a new Cloud Run MCP server, which empowers MCP-compatible AI agents (like AI assistants, IDE integrations, or SDKs) to programmatically deploy applications. Read more here.

The news didn’t stop with I/O – we announced several important announcements to help you deploy AI at scale:

  1. Introducing the next generation of AI inference, powered by llm-d: We’re making inference even easier and more cost-effective, by making vLLM fully scalable with Kubernetes-native distributed and disaggregated inference. This new project is called llm-d. Google Cloud is a founding contributor alongside Red Hat, IBM Research, NVIDIA, and CoreWeave, joined by other industry leaders AMD, Cisco, Hugging Face, Intel, Lambda, and Mistral AI.
  2. Mistral AI's Le Chat Enterprise and Mistral OCR 25.05 model are available on Google Cloud. Available today on Google Cloud Marketplace, Mistral AI's Le Chat Enterprise is a generative AI work assistant designed to connect tools and data in a unified platform for enhanced productivity.
  3. Anthropic’s Claude Opus 4 and Claude Sonnet 4 on Vertex AI. Claude Opus 4 and Claude Sonnet 4 are generally available as a Model-as-a-Service (MaaS) offering on Vertex AI. For more information on the newest Claude models, visit Anthropic’s blog.

…and made strides in security:

  1. What’s new with Google Cloud’s Risk Protection Program. We unveiled at Google Cloud Next major updates to our Risk Protection Program, an industry-first collaboration between Google and insurers that provides competitively priced cyber-insurance and broad coverage for Google Cloud customers. We’re now including Affirmative AI insurance coverage for your Google-related AI workloads. Here’s what’s new.
  2. How Confidential Computing lays the foundation for trusted AI. Our latest Confidential Computing innovations highlight the creative ways our customers are using Confidential Computing to protect their most sensitive workloads — including AI.
  3. How governments can use AI to improve threat detection and reduce cost. In the latest Cloud CISO Perspectives newsletter, our Office of the CISO’s Enrique Alvarez, public sector advisor, explains how government agencies can use AI to improve threat detection — and save money.

News you can use: Actionable ways to get started

Get fluent in generative AI:

62% of employers now expect candidates and employees to possess at least some familiarity with AI. That’s why we launched a first-of-its kind generative AI certification for non-technical learners—plus a new suite of no-cost training to help you prepare for that certification. That means you — and your company — can be among the first to take advantage of this opportunity to validate your strategic acumen in gen AI. Become a generative AI leader today.

Then, put generative AI to work

At I/O, we expanded generative AI media on Vertex AI. But how do you get started, today? To help you make the most of all the latest generative AI media announcements, we redesigned Vertex AI Studio. The developer-first experience will be your source for generative AI media models across all modalities. You’ll have access to Google’s powerful generative AI media models such as Veo, Imagen, Chirp and Lyria in the Vertex AI Media Studio.

https://storage.googleapis.com/gweb-cloudblog-publish/original_images/1-overview-helpmewrite_c5csojo.gif

Redesigned Vertex AI Studio

To help you turn your generative AI ideas into real web applications, we published this guide to create gen AI apps in less than 30 seconds with Vertex AI and Cloud Run. Any developer knows it’s a complex process to build shareable, interactive applications: you have to set up infrastructure, wire APIs, and build a front-end. It's usually a complex process. What if you could skip the heavy lifting and turn your generative AI concept into a working web app with just a few clicks?

New how-to series alert: Text-to-SQL agents

Recently, powerful large language models (LLMs) like Gemini, with their abilities to reason and synthesize, have driven remarkable advancements in the field of text-to-SQL. In this blog post, the first entry in a series, we explore the technical internals of Google Cloud's text-to-SQL agents.

Real-life demo: What if we turned Gemini into an AI basketball coach?

We rounded out this month with a deep-dive into a demo we showcased at Google Cloud Next and most recently, at I/O. In this article, we showed an AI experiment that turns Gemini 2.5 Pro into a jump shot coach. By combining a ring of Pixel cameras with Vertex AI, the coaching system connects AI motion capture, biomechanical analytics, and Gemini-powered coaching via text and voice.

“It’s like we always say: AI is only as good as the information you give it. For the AI basketball coach to be accurate, we knew we had to talk to actual, real-life professionals. So we talked to our partners at the Golden State Warriors and came up with essential criteria for helping you shoot like the pros.”

Stay tuned for monthly updates on Google Cloud’s AI announcements, news, and best practices. For a deeper dive into the latest from Google Cloud, read our weekly updates, The Overwhelmed Person’s Guide to Google Cloud.


April

From Agent2Agent Protocol to Gemini 2.5 Pro, it’s been an exciting month. We hosted Google Cloud Next in Las Vegas on April 9th with over 30,000 people, announcing incredible innovations from Ironwood TPU to Agent2Agent Protocol. We also expanded Vertex AI in dizzying ways – now, it’s the only platform with generative media models across video, image, speech, and music and makes it possible for multiple agents to coexist in your enterprise.

If you missed the livestream, take a look at our Day 1 recap and summary of the developer keynote. It’s been incredible to see how customers have been applying AI to hundreds of use cases. In fact, we’ve counted more than 600 examples.

Top announcements

Agents

Our recently launched Agent2Agent (A2A) protocol is getting a lot of attention, and for good reason. This open interoperability protocol is designed to make it easy for AI agents to communicate with one another, no matter its foundation. It’s also a powerful complement to Anthropic’s Model Context Protocol (MCP). Together, these two open protocols are the foundation for complex multi-agent systems.

Since the launch, we’ve been excited to see all the discussions and reactions across social media and YouTube. The A2A GitHub repository (13k stars) has grown quickly, which underscores the industry's eagerness for a standardized agent communication protocol, especially these features:

And, speaking of agents, we made a host of updates to Google Agentspace – starting with unified search. Now,Agentspace is integrated with Chrome Enterprise, which means you can search and access all of your enterprise’s resources — right from the search box in Chrome. You can also discover and adopt agents quickly and easily with Agent Gallery, and create agents with our new no-code Agent Designer.

Models

This month, we announced six new models (some in preview, some generally available) for our customers:

Security

Hot on the heels of our AI Protection news, we introduced Google Unified Security at Next ‘25, which lays the foundation for superior security outcomes, followed by the RSA Conference in San Francisco. We showcased there:

News you can use:

In just the last year alone, we’ve seen over 40x growth in Gemini use on Vertex AI, now with billions of API calls per month. So what’s new and better with Vertex AI?

To start:

Meta’s Llama 4 is generally available on Vertex AI.

Worth a second read:

Next month, we’ll be releasing several guides and deep-dives into how to use all these features, so stay tuned!

Hear from our leaders:

As usual, we ended this month with our monthly column, The Prompt. In this installment, we hear from Logan Kilpatrick, AI product leader on Google DeepMind, on how multimodal capabilities – especially audio and vision – are bringing about a new UX paradigm. Here’s a snippet:

“There’s the age-old quote that a picture is worth a thousand words. This matters double in the world of multimodal. If I look at my computer right now and try to describe everything I see, it would take 45 minutes. Or, I could just take a picture. Use cases for vision could be something from as simple as object tracking to image detection. For example, a factory watching an assembly line to make sure there's no impurities in the product they’re creating. Or analyzing dozens of pictures of your farm and trying to understand crop yields. There's huge breadth and opportunity by blending these modalities together.”

Stay tuned for monthly updates on Google Cloud’s AI announcements, news, and best practices. For a deeper dive into the latest from Google Cloud, read our weekly updates, The Overwhelmed Person’s Guide to Google Cloud.


March

In March, we made big strides in our models and accessibility – from announcing Gemini 2.5, our most intelligent AI model, to Gemma 3, our most capable and advanced version of the Gemma open-model family. We also introduced Gemini Code Assist for individuals, a free version of our AI coding assistant.

In the industry, we shone light on telecom and gaming. We went to Mobile World Congress, where we displayed key agent use cases where AI is becoming an imperative in telecom. At the same time, we talked to leaders in the games industry, who told us how they’re using Google Cloud AI to drive unprecedented advancements in game development, including smarter, faster, and more immersive gaming experiences.

Top announcements

The Gemini family is growing – we introduced Gemini 2.5, a thinking model designed to tackle increasingly complex problems. Gemini 2.5 Pro Experimental leads common benchmarks by meaningful margins and showcases strong reasoning and code capabilities. Learn all about it here.

We also introduced Gemma 3, the most capable model you can run on a TPU or GPU. To help you get started, we shared guides on how to deploy your AI workloads on Cloud Run and Vertex AI.

In the world of open-source, we announced Claude 3.7 Sonnet, Anthropic’s most intelligent model to date and the first hybrid reasoning model on the market, is available in preview on Vertex AI Model Garden. Claude 3.7 Sonnet can produce quick responses or extended, step-by-step thinking that is made visible to the user. Explore our sample notebook and documentation to start building.

Finally, we took a step forward in security. We introduced AI Protection, a set of capabilities designed to safeguard AI workloads and data across clouds and models — irrespective of the platforms you choose to use. Our Office of the CISO suggests these five do-it-today, actionable tips on how leaders can help their organizations adopt AI securely.

News you can use

Wondering how to make the most of your AI? When it comes to Gemini 2.0, we created a helpful guide that teaches you how to optimize one of enterprises’ most time-intensive tasks: document extraction.

From an infrastructure perspective, we broke down the four top use cases for AI Hypercomputer and ways to get started. In this guide, you’ll learn everything from affordable inference to reducing delivery bottlenecks. Take Moloco, for example. Using the AI Hypercomputer architecture they achieved 10x faster model training times and reduced costs by 2-4x.

And, speaking of cost – do you know the true cost of enterprise AI? Enterprises need ways to optimize large AI workloads because these resources can still be quite expensive. Learn how to calculate your AI cost and dig into these five tips to optimize your workflow on Google Cloud Platform.

Hear from our leaders

We closed off this month with our monthly column, The Prompt. In this installment, we hear from Suraj Poozhiyil, AI product leader, on how AI agents depend on “enterprise truth” – your enterprise’s unique context – to be successful. Here’s a snippet:

We’ve heard it before – AI is only as good as the data you put into it. When it comes to agents, AI agents are only as good as the context you give them. Enterprise truth is the answer to questions like, "What is our company’s policy for creating a purchase order?" and "What is the specific workflow for approvals and compliance?"

Stay tuned for monthly updates on Google Cloud’s AI announcements, news, and best practices. For a deeper dive into the latest from Google Cloud, read our weekly updates, The Overwhelmed Person’s Guide to Google Cloud.


February

2025 is off to a racing start. From announcing strides in the new Gemini 2.0 model family to retailers accelerating with Cloud AI, we spent January investing in our partner ecosystem, open-source, and ways to make AI more useful. We’ve heard from people everywhere, from developers to CMOs, about the pressure to adapt the latest in AI with efficiency and speed – and the delicate balance of being both conservative and forward-thinking. We’re here to help. Each month, we’ll post a retrospective that recaps Google Cloud’s latest announcements in AI – and importantly, how to make the most of these innovations.

Top announcements: Bringing AI to you

This month, we announced agent evaluation in Vertex AI. A surprise to nobody, AI agents are top of mind for many industries looking to deploy their AI and boost productivity. But closing the gap between impressive model demos and real-world performance is crucial for successfully deploying generative AI. That’s why we announced Vertex AI’s RAG Engine, a fully managed service that helps you build and deploy RAG implementations with your data and methods. Together, these new innovations can help you build reliable, trustworthy models.

From an infrastructure perspective, we announced new updates to AI Hypercomputer. We wanted to make it easier for you to run large multi-node workloads on GPUs by launching A3 Ultra VMs and Hypercompute Cluster, our new highly scalable clustering system. This builds on multiple advancements in AI infrastructure, including Trillium, our sixth-generation TPU.

What’s new in partners and open-source

This month, we invested in our relationship with our partners. We shared how Gemini-powered content creation in Partner Marketing Studio will help partners co-market faster. These features are designed to streamline marketing efforts across our entire ecosystem, empowering our partners to unlock new levels of success, efficiency, and impact.

At the same time, we shared several important announcements in the world of open-source. We announced Mistral AI’s Mistral Large 24.11 and Codestral 25.01 models on Vertex AI. These models will help developers write code and build faster – from high-complexity tasks to reasoning tasks, like creative writing. To help you get started, we provided sample code and documentation.

And, most recently, we announced the public beta of Gen AI Toolbox for Databases in partnership with LangChain, the leading orchestration framework for developers building LLM applications. Toolbox is an open-source server that empowers application developers to connect production-grade, agent-based generative AI applications to databases. You can get started here.

Industry news: Google Cloud at the National Retail Federation (NRF)

The National Retail Federation kicked off the year with their annual NRF conference, where Google Cloud showed how AI agents and AI-powered search are already helping retailers operate more efficiently, create personalized shopping experiences, and use AI to get the latest products and experiences to their customers. Check our new AI tools to help retailers build gen AI search and agents.

As an example, Google Cloud worked with NVIDIA to empower retailers to boost their customer engagements in exciting new ways, deliver more hyper-personalized recommendations, and build their own AI applications and agents. Now with NVIDIA's AI Enterprise software available on Google Cloud, retailers can handle more data and more complex AI tasks without their systems getting bogged down.

News you can use

This month, we shared several ways to better implement fast-moving AI, from a comprehensive guide on Supervised Fine Tuning (SFT), to how developers can help their LLMs deliver more accurate, relevant, and contextually aware responses, minimizing hallucinations and building trust in AI applications by optimizing their RAG retrieval.

We also published new documentation to use open models in Vertex AI Studio. Model selection isn’t limited to Google’s Gemini anymore. Now, choose models from Anthropic, Meta, and more when writing or comparing prompts.

Hear from our leaders

We closed out the month with The Prompt, our monthly column that brings observations from the field of AI. This month, we heard from Warren Barkley, AI product leader, who shares some best practices and essential guidance to help organizations successfully move AI pilots to production. Here’s a snippet:

More than 60% of enterprises are now actively using gen AI in production, helping to boost productivity and business growth, bolster security, and improve user experiences. In the last year alone, we witnessed a staggering 36x increase in Gemini API usage and a nearly 5x increase of Imagen API usage on Vertex AI — clear evidence that our customers are making the move towards bringing gen AI to their real-world applications.

Stay tuned for monthly updates on Google Cloud’s AI announcements, news, and best practices. For a deeper dive into the latest from Google Cloud, read our weekly updates, The Overwhelmed Person’s Guide to Google Cloud.

Posted in