Google I/O 2025: From research to reality

At Google I/O, they  showed Google is fully in the Gemini era, showcasing significant AI innovations across products, research, and infrastructure, all geared towards making AI more helpful for everyone. Check out the recap below of what to expect.

From research to reality — we’re taking cutting-edge research to build the most helpful AI that’s more intelligent, agentic and personalized. 

And we’re building on impressive momentum — every one of our 15 products with over half a billion users now uses Gemini, our natively multimodal AI model family. 

Intelligent
We are shipping models at a relentless pace — in March, we released Gemini 2.5 Pro, setting new state-of-the-art across key benchmarks and representing a significant leap in reasoning, coding, science and maths capabilities, and we’ve built on that progress today: 

  • Gemini 2.5 advancements: Gemini 2.5 Flash is now our default model, blending incredible quality with lightning-fast response times. Our enhanced reasoning model, Deep Think (in Gemini 2.5 Pro) uses our latest research on thinking and reasoning 

  • New creative tools: Introducing Veo 3 and Imagen 4, our frontier-pushing video and image generation models. Creatives: try Flow, our new AI filmmaking tool, to create story elements (cast, locations, objects, styles) with natural language, all in one place.

 

Agentic
By applying our cutting-edge AI models across Google products, we can make them more agentic — capable of intelligently taking action on your behalf, with your permission, across a variety of tasks, from complex academic research to restaurant reservations, across the web and in your Google apps:

 

  • Project Mariner: Launched last December, Project Mariner now includes a system of agents that can complete up to ten different tasks at a time. Starting in the US, these agents can simultaneously help you look up information, make bookings, buy things, do research and more

  • Get things done, directly from Search: We’re building the agentic capabilities of Project Mariner into AI Mode in the US, helping people save time with tasks like purchasing tickets to a baseball game

  • Project Astra in Google Search: We’re integrating Project Astra’s Live capabilities into Search in the US, so you can talk back-and-forth with Search about what you see in real-time, using your camera. Stumped on a project? Simply tap the “Live” icon in AI Mode, point your camera, ask your question, and get an explanation, with links to other resources you can explore in more detail 

  • Shopping features: Our new shopping experience in AI Mode combines Gemini model capabilities with our Shopping Graph to help you browse for inspiration, and choose products that are the “right fit”. See how an outfit looks on you, and ask our agentic checkout feature to make the purchase when the price is right. 

 

Personalised
Combine knowledge of the real world with information you share about your world, so AI can be even more helpful. Optionally share personal context like search history, emails, or your schedule allows for AI responses tailored to your preferences:

  • Gemini: Our goal for the Gemini app is to become the most helpful universal AI assistant. We’re enhancing Gemini Live in the Gemini app with personal context. With your permission, it will integrate with Google services like Maps, Calendar, Tasks, and Keep (and more later) for deeper daily assistance. And it’s getting even better with more Project Astra live capabilities coming soon to Gemini Live. 

  • AI Mode: our most powerful AI search, features advanced reasoning, multimodality, and deeper exploration via follow-up questions and web links. It now also offers personalized suggestions based on past searches or other Google apps, starting with Gmail. Search “Nashville weekend, friends, food & music,” and AI Mode suggests restaurants from your booking history and activities near your hotel.

  • Gemini Deep Research & Canvas: Make Deep Research personally powerful by blending your own private files with public data for insights unique to your context. Then, let Canvas bring your personal vision to life, as Gemini 2.5 Pro turns your descriptions into working code, apps, and creative content faster than ever.

 

With this, and all our groundbreaking work, we’re building AI that’s more intelligent, agentic and personalised, enriching our lives, advancing the pace of scientific progress and ushering in a new age of discovery and wonder.

See our Google Reviews

Google I/O wrap up note

A year ago on the I/O stage Goolge first shared  plans for Gemini, a family of natively multimodal AI models that could reason across text, images, video, code, and more. When we launched the Gemini era in December 2023, it marked a big step in turning any input into any output — an “I/O” for a new generation.

Google I/O showed how Google is fully in the Gemini era, showcasing AI innovations across products, research, and infrastructure, and how it brings us close to our ultimate goal of making AI helpful for everyone. Sundar Blog Post.

All Google products with more than two billion users are built with Gemini. Today we shared how that helps us to create new experiences and make our products even more helpful:

  • Expanding AI Overviews in Search. With a new customised Gemini model — capable of multi-step reasoning, planning and multimodality — combined with best-in-class Search you’ll soon be able to ask complex, multi-step questions, customise search results, and even ask questions with videos [Blog Post]

  • Introducing Ask Photos. Over six billion photos are uploaded every day to Google Photos. With Gemini’s multimodal capabilities, we’re redefining how you can search your photos and videos. Want to find a specific memory or recall information hidden in your gallery? Just Ask Photos. [Blog Post]

  • New ways to engage with Gemini in Workspace. Gemini’s capabilities will expand to more users and integrate into the side panel of Gmail, Docs, Drive, Slides, and Sheets. Gemini features will also be added to the Gmail mobile app. [Blog Post]

  • Gemini for Android. We’re building AI right into the Android operating system. Students can now get homework help by circling problems with Circle to Search. And Gemini’s overlay will provide dynamic suggestions related to what’s on your screen — summarise a PDF or “ask this video” — while TalkBack with Gemini will be capable of even more detailed image descriptions. [Blog Post]

We’re also bringing Gemini 1.5 Pro to Gemini Advanced subscribers in more than 35 languages, along with a 1 million token context window — the longest of any widely available consumer chatbot in the world. This means it can understand more information than ever before,  like a 1500-page PDF and soon, 30,000 lines of code and an hour-long video.

  • Gemini Advanced subscribers will also soon get access to Live, a new mobile conversational experience. With Live, you can talk to Gemini and choose from different natural-sounding voices. You can speak at your own pace and even interrupt with questions, making conversations more intuitive. [Blog Post]

All of this work is underpinned by our technical leadership in building the world’s most advanced AI.

And we’re looking ahead too, to the next models: we shared more details about Gemini 1.5 Flash, a more cost-efficient model built based on user feedback, with lower latencies; and Project Astra, our vision for the next generation of AI assistants, a responsive agent that can understand and react to the context of conversations. [Blog Post]

We’ve also been working closely with the creative community to explore how generative AI can best support the creative process, and to make sure our AI tools are as useful as possible at each stage:

  • Today, we’re introducing Veo, our most capable model for generating high-definition video, and Imagen 3, our highest quality text-to-image model. We’re also sharing new demo recordings — with global artists — created with our Music AI Sandbox. [Blog Post]

Of course, these advancements in AI are only made possible by truly cutting-edge infrastructure technology. Training state-of-the-art models requires a lot of computing power.

  • Today we unveiled the 6th generation of our TPUs, called Trillium, which delivers a 4.7x improvement in compute performance per chip over the previous generation, TPU v5e, and which we’ll make available to Cloud customers later this year. [Blog Post]

Bold innovation must be underpinned by responsible innovation. So we’re developing a cutting-edge technique we call AI-assisted red teaming, that draws on Google DeepMind’s gaming breakthroughs like AlphaGo and expanding our technical watermarking innovations like SynthID in two new modalities — text and video —  so AI-generated content is easier to identify. [Blog Post]

By using the power of Gemini, we plan to make AI useful for everyone. Google’s mission is to organise the world’s information across every input, making it accessible via any output, and combining the world’s information, with the information in YOUR world, in a way that’s truly useful for you. Gemini will help us towards that goal.