AI Mode, Veo 3, Imagen 4, Android XR, and More

[ad_1]

Google’s annual event I/O has returned this year, pushing the boundaries of AI further than ever before. What started off with the keynote speech by Google CEO Sundar Pichai, highlighting the milestones accomplished by the tech giant, soon escalated to an ever-exciting show of AI-powered advancements and new generative AI tools. From the new AI mode on Google Search and Gemini Live to the launch of Veo 3, Imagen 4, and Flow, to the unveiling of Android XR and Samsung Moohan – Google was pulling one AI rabbit after the other out of the hat. Of all that was said and shown, this blog brings you the 8 biggest AI breakthroughs and launches announced at the Google I/O 2025 event.

1. Google Beam & Real-Time Translation in Google Meet

Google has taken video calling to a whole new level with Google Beam – an evolution of Project Starline that offers immersive 3D video communication. This new technology captures the views of the speaker from 6 different camera angles, along with their movement at 60 fps. It then puts them all together to generate a 3D version of the speaker, making it feel like the person is right in front of you. Aimed at making virtual interactions feel more lifelike, Google Beam will soon be available to Google Meet users in the US and then to other countries.

Complementing this, Google Meet now features real-time speech translation. Powered by AI, this translation feature can pick up your dialect, tone, and nuances to give accurate translations in real-time. Initially supporting English and Spanish, Google plans to add more languages soon, facilitating seamless multilingual conversations during video calls. This new feature has already been rolled out to US users and will soon be launched worldwide. Google Enterprise users will also get access to this feature towards the end of this year.

The biggest announcement made at Google I/O 2025 has got to be about the new AI Mode in Google Search. Owing to the widespread acceptance of AI overviews on Google Search, they have now brought the power of AI directly to the search bar with the AI Mode. This new feature lets users use AI directly to search for results, just as they would on ChatGPT, Gemini, or any other AI chatbot.

With an expanded search window, users can now add more context and ask multiple questions within the same search query. Google Search breaks user queries into multiple smaller queries and categories and runs parallel searches on all of them. With AI-powered reasoning capabilities, it then puts together all the info and generates a comprehensive and contextual response. This transforms Google Search into a more interactive experience.

Key Features

Google Search’s new AI Mode offers 7 new features:

  1. Personal Context: You can now get Google to give you personalized responses by integrating your search history and data from other Google apps and tools like Gmail. This integration lets the AI understand your style and choices, to generate smarter responses that are uniquely helpful to you.
  2. Deep Research: This feature multiples the web search capabilities of Google to do dozens or even hundreds of searches at the same time, to gather more information, resulting in more detailed and well-researched responses.
  3. Multiple Response Formats: The now AI-powered Google Search dynamically generates the best layout for each response, based on the query. For instance, it can intelligently generate interactive lists and graphs for sports and financial queries.
  4. Personalized Shopping Suggestions: Instead of simply listing out product pages and shopping links, Google Search can now give you personalized shopping suggestions based on your taste, previous searches, and purchase history. While you can add more context and details to your search query, Google also recommends points to consider to help you make the right choice.
  5. Virtual Outfit Trials: Another highlight of the AI Mode is the AI-powered shopping with virtual try-ons. You can now virtually try on clothes before buying them, directly on Google Search. Simply select the outfit, upload your image, and watch as Google magically dresses you up in that outfit right on the screen. This feature has also been rolled out to users in the US today.
  6. Search Live: You can now do live video calls to Google Search for real-time visual assistance, similar to the Gemini Live feature on the chatbot.
  7. AI-powered Visual Search: Where Google Lens would earlier find similar images based on the input image, it can now give AI overviews of any image you click or upload. It can basically explain anything that’s in front of your eyes, being a virtual companion, especially to the visually impaired.

The AI Mode on Google Search is currently being rolled out to US users. Google plans to make it available in other countries soon.

3. Project Astra: Gemini Live

The biggest update of the Gemini chatbot announced at the Google I/O event this year was the Gemini Live feature. An extension of Google’s Project Astra, Gemini Live is developed to be a universal AI assistant. It lets users have live video calls with the AI-powered Gemini chatbot, offering real-time AI assistance for anything and everything. It lets users engage in interactive camera conversations, receive on-the-go translations, and share screens or camera feeds for help. This feature is now available in over 45 languages across 150+ countries, to both Android and iOS users.

4. Project Mariner: Agent Mode

At Google I/O 2025, the company demoed Agent Mode – a Project Mariner-based AI agent with computer use features. This superagent is capable of doing up to 10 tasks at once. It can make calls, search the web, find YouTube videos, give you suggestions, answer questions, and do a lot more. It is also intelligent enough to learn the workflow from one task and apply it to do other tasks similarly, using a technology called ‘teach and repeat’.

Agent Mode is designed to be personal, proactive, and powerful. It can access your calendar, see upcoming events, and set reminders or prep you for the event, even before you ask it to. This level of autonomy and intelligence hasn’t been seen in general-use AI agents before. It helps in automating a number of everyday tasks like scheduling, notetaking, interview prep, etc.

Agent Mode in Action

Google took it even a step further by integrating it with Google Search’s AI Mode to bring its users an agentic search feature. With this, users can run multiple web searches and web search-based tasks in the background, which will be done autonomously by the agent.

For example you can use this feature to set up agentic checkout for e-shopping. Once you find a product that you wish to buy, you can use agentic search to find it for you within your budget. Google Search would keep tracking the price across websites, and once the price drops to your range, it will automatically place your order with just one tap. You can even pay for it using Google Pay, again with just a tap.

So now, just like Agent2Agent Protocol, and Anthropic’s Model Context Protocol, Gemini API and SDK will be able to use MCP tools. Google will soon roll out Project Mariner’s computer use capabilities to developers via the Gemini API. Meanwhile, an experimental version of the multi-tasking Agent Mode is now available to Google AI Ultra subscribers in the US.

Google announced some of its latest and most advanced generative AI tools at the Google I/O 2025 event. This included:

  1. Music AI Sandbox with Lyria 2: The Music AI Sandbox, powered by Lyria 2, enables users to generate music compositions using AI. It can create harmonies, rhythms, background scores, and even full compositions with orchestra based on user input.
  2. Genie 2: This advanced tool from Google can transform 2D Images into interactive 3D environments in just 2 steps and a prompt. It has a wide range of applications in gaming, virtual reality, and digital content creation.
  3. Imagen 4: Imagen 4 is Google’s latest text-to-image generation model, capable of producing high-quality, photorealistic images from textual descriptions. Not only does it get text and spelling right, but it can also intelligently select the right font, font size, etc., based on the query. Moreover, it works up to 10x faster than previous models.
  4. Veo 3: Google launched its latest version of Veo at the annual event. The upgraded Veo 3 takes AI-powered video generation to a whole new level, creating hyper-realistic and high-quality videos from text prompts. Along with video, it also generates realistic audio output including dialogues and background sounds.

  1. Flow: This new filmmaking tool from Google brings together the creative capabilities of Veo, Imagen, and Gemini. It allows users to generate short films from text or image prompts, integrating sound, dialogue, and visual effects. With text-to-image, image-to-video, and text-to-video features, it becomes a one-stop shop for bringing imagination to reality. Moreover, it also comes with scene extension and editing features.

These advanced tools are now available to Google AI Pro and Ultra plan subscribers and will slowly be integrated into the Google Gemini chatbot.

6. Gemini App Integration with Imagen 4, Veo 3, and More

Google I/O 2025 was more about Gemini than about AI, as proven by CEO Sundar Pichai’s word counter. Several announcements regarding Google’s Gemini chatbot were made at the event including updates on Deep Research and Canvas and integrations with Google’s latest generative AI tools.

Gemini Updates Launched at Google I/O 2025

Google Gemini Updates: AI Search, Flow, Veo

Here’s a list of all the Gemini Updates revealed at the Google I/O event this year.

  1. Gemini in Chrome: The next big thing is that Google will soon be rolling out Gemini on Google Chrome as a web browsing AI agent. This lets users ask their search queries and follow-up questions about the search results, directly to the AI chatbot.

  1. Gemini Voice: Google has integrated native audio output into Gemini’s Voice Mode which lets it respond to users in a more personalized and nuanced manner. It can switch between languages, change tones, and even whisper during the same conversation. You can test out this updated version via the Gemini API.
  2. Deep Research: You can now upload your own files to guide the research agent while doing Deep Research using Google Gemini. You can also connect it to your Gmail and Google Drive to fetch more data or provide some context.

  1. Canvas: The Canvas feature on Gemini can now convert deep research reports into custom podcasts, quizzes, infographics, and more.

  1. Imagen 4: Google Gemini’s image generation capabilities are now powered by Imagen 4 making the images more realistic and detailed.
  2. Veo 3: Gemini can now generate realistic videos with accurate audio, dialogues, and background sound, thanks to the newly integrated Veo 3.

These updates will be rolled out to subscribers in the coming weeks.

7. Android XR and Samsung Moohan

Android XR is Google’s first-ever Android platform, venturing into extended reality. This technology, powered by Gemini, fosters an immersive experience for users through hyper-realistic videos in real-time. Samsung’s Moohan, a newly designed pair of smart glasses, would be the first device that leverages Android XR for AI assistance. These glasses offer features like real-time navigation, translation, and camera live-streaming, aiming to enhance user interaction with the digital world.

With these glasses you can watch live events from your home, as if you were sitting in the front row of the stadium. With the ability to show Google Maps in 3D, it can visually take you places in real-time, giving you a realistic experience. Moreover, it comes with memory and can answer questions. Designed to provide AI assistance in real-time just like a human companion, Samsung Moohan can click pictures, make bookings, and even translate audio to text. Unlike most other smart glasses that come in a single sci-fi inspired design, these ones are going to be designed in various styles by Gentle Monster and Warby Parker.

8. New Google AI Subscription Models

Apart from all these launches and updates, Google also introduced two new subscription plans at its annual event:

  • AI Pro: Priced at $19.99/month, this plan offers users all its advanced AI features with high limits. This is suitable for general users worldwide.
  • AI Ultra: Priced at $249/month, this plan gives access to the most advanced AI tools and models, including experimental features, with higher rate limits. This is targeted at professionals and enterprises in the US only.
Google AI Subscription Pricing

Conclusion

Google I/O 2025 was quite the show, giving us all a glimpse into the ambitious AI plans Google has. From enhancing everyday tools like Google Search and Google Meet to developing advanced creative tools like Flow and Genie 2, Google’s innovations aim to redefine the boundaries of AI. As these updates and models roll out, I’m sure AI will be an integral part of a common man’s everyday life. Be it Project Astra, Project Mariner, or Android XR – these developments mark a significant step toward a more intuitive and immersive digital future, powered by AI.

K.C. Sabreena Basheer

Sabreena is a GenAI enthusiast and tech editor who’s passionate about documenting the latest advancements that shape the world. She’s currently exploring the world of AI and Data Science as the Manager of Content & Growth at Analytics Vidhya.

Login to continue reading and enjoy expert-curated content.

[ad_2]

Leave a Comment