News from the AI & ML world

DeeperML - #googleio

Ellie Ramirez-Camara@Data Phoenix //
Google has recently unveiled a suite of advancements in its AI media generation models at Google I/O 2025, signaling a major leap forward in the field. The highlights include the launch of Veo 3, the first video generation model from Google with integrated audio capabilities, alongside Imagen 4, and Flow, an AI filmmaking tool. These new tools and upgrades to Veo 2 are designed to provide creators with enhanced realism, emotional nuance, and coherence in AI-generated content. These upgrades are designed to target professional markets and are available to Ultra subscribers via the Gemini app and Flow platform.

The most notable announcement was Veo 3, which allows users to generate videos with synchronized audio, including ambient sounds, dialogue, and environmental noise. This model understands complex prompts, enabling users to create short stories brought to life with realistic physics and accurate lip-syncing. Veo 2 also received significant updates, including the ability to use images as references for character and scene consistency, precise camera controls, outpainting capabilities, and object manipulation tools. These enhanced features for Veo 2 are aimed at providing filmmakers with greater creative control.

Also introduced was Flow, an AI-powered video creation tool that integrates the Veo, Imagen, and Gemini models into a comprehensive platform. Flow allows creators to manage story elements such as cast, locations, objects, and styles in one interface, enabling them to combine reference media with natural language narratives to generate scenes. Google also introduced "AI Mode" in Google Search and Jules, a powerful new asynchronous coding agent. These advancements are part of Google's broader effort to lead in AI innovation, targeting professional markets with sophisticated tools that simplify the creation of high-quality media content.

Recommended read:
References :
  • Data Phoenix: Google announced several updates across its media generation models
  • pub.towardsai.net: TAI #154: Gemini Deep Think, Veo 3’s Audio Breakthrough, & Claude 4’s Blackmail Drama
  • THE DECODER: Google says Veo 3 users have generated millions of AI videos in just a few days

Aminu Abdullahi@eWEEK //
Google has unveiled significant advancements in its AI-driven media generation capabilities at Google I/O 2025, showcasing updates to Veo, Imagen, and Flow. The updates highlight Google's commitment to pushing the boundaries of AI in video and image creation, providing creators with new and powerful tools. A key highlight is the introduction of Veo 3, the first video generation model with integrated audio capabilities, addressing a significant challenge in AI-generated media by enabling synchronized audio creation for videos.

Veo 3 allows users to generate high-quality visuals with synchronized audio, including ambient sounds, dialogue, and environmental noise. According to Google, the model excels at understanding complex prompts, bringing short stories to life in video format with realistic physics and accurate lip-syncing. Veo 3 is currently available to Ultra subscribers in the US through the Gemini app and Flow platform, as well as to enterprise users via Vertex AI, demonstrating Google’s intent to democratize AI-driven content creation across different user segments.

In addition to Veo 3, Google has launched Imagen 4 and Flow, an AI filmmaking tool, alongside major updates to Veo 2. Veo 2 is receiving enhancements with filmmaker-focused features, including the use of images as references for character and scene consistency, precise camera controls, outpainting capabilities, and object manipulation tools. Flow integrates the Veo, Imagen, and Gemini models into a comprehensive platform allowing creators to manage story elements and create content with natural language narratives, making it easier than ever to bring creative visions to life.

Recommended read:
References :
  • Data Phoenix: Google updated its model lineup and introduced a 'Deep Think' reasoning mode for Gemini 2.5 Pro
  • Maginative: Google’s revamped Canvas, powered by the Gemini 2.5 Pro model, lets you turn ideas into apps, quizzes, podcasts, and visuals in seconds—no code required.
  • Replicate's blog: Generate incredible images with Google's Imagen-4
  • AI News | VentureBeat: At Google I/O, Sergey Brin makes surprise appearance — and declares Google will build the first AGI
  • www.tomsguide.com: I just tried Google’s smart glasses built on Android XR — and Gemini is the killer feature
  • Data Phoenix: Google has launched major Gemini updates, including free visual assistance via Gemini Live, new subscription tiers starting at $19.99/month, advanced creative tools like Veo 3 for video generation with native audio, and an upcoming autonomous Agent Mode for complex task management.
  • sites.libsyn.com: Google's VEO 3 Is Next Gen AI Video, Gemini Crushes at Google I/O & OpenAI's Big Bet on Jony Ive
  • eWEEK: Google’s Co-Founder in Office ‘Pretty Much Every Day’ to Work on AI
  • learn.aisingapore.org: Advancing Gemini’s security safeguards – Google DeepMind
  • Google DeepMind Blog: Gemini 2.5: Our most intelligent models are getting even better
  • TestingCatalog: Opus 4 outperforms GPT-4.1 and Gemini 2.5 Pro in coding benchmarks
  • LearnAI: Updates to Gemini 2.5 from Google DeepMind
  • pub.towardsai.net: This week, Google’s flagship I/O 2025 conference and Anthropic’s Claude 4 release delivered further advancements in AI reasoning, multimodal and coding capabilities, and somewhat alarming safety testing results.
  • learn.aisingapore.org: Updates to Gemini 2.5 from Google DeepMind
  • Data Phoenix: Google announced several updates across its media generation models
  • thezvi.wordpress.com: Fun With Veo 3 and Media Generation
  • Maginative: Google Gemini Can Now Watch Your Videos on Google Drive
  • www.marktechpost.com: A Coding Guide for Building a Self-Improving AI Agent Using Google’s Gemini API with Intelligent Adaptation Features

Hamish Hector@techradar.com //
Google is making a renewed push into the smart glasses market with the upcoming Android XR glasses, leveraging its Gemini AI to enhance user experience. During the I/O 2025 developer conference, Google showcased the capabilities of these glasses, highlighting features such as live language translation. The Android XR glasses can connect to a smartphone to access apps and come equipped with speakers. An optional in-lens display will allow users to view information privately.

Google is partnering with eyewear brands such as Warby Parker and Gentle Monster, focusing on creating stylish and wearable designs. These collaborations aim to move away from the bulky, tech-heavy aesthetic often associated with earlier smart glasses models. This partnership hints that Google is taking style a lot more seriously this time around. Warby Parker is well known as a direct-to-consumer eyewear brand that makes it easy to get trendy glasses at a relatively accessible price. Meanwhile, Gentle Monster is currently one of the buzziest eyewear brands that isn’t owned by EssilorLuxottica. The Korean brand is popular among Gen Z, thanks in part to its edgy silhouettes and the fact that Gentle Monster is favored by fashion-forward celebrities like Kendrick Lamar, Beyoncé, Rihanna, Gigi Hadid, and Billie Eilish. Partnering with both brands seems to hint that Android XR is aimed at both versatile, everyday glasses as well as bolder, trendsetting options.

The glasses integrate Google’s Gemini AI assistant, enabling users to interact with their surroundings and access information hands-free. Google is working with Xreal and Qualcomm on Project Aura, an optical see-through XR device, to integrate Android XR software. Google also plans to integrate Project Astra to allow users to talk back and forth with Search about what they see in real time with their camera. The tech giant also rolled out AI Mode to Google Search for every U.S. user in order to answer your question before showing the traditional list of links.

Recommended read:
References :
  • www.techradar.com: Xreal is making Android XR glasses and this could be the biggest XR announcement since the Meta Quest 2 – here’s why
  • www.engadget.com: Google demos Android XR glasses at I/O with live language translation
  • techvro.com: Google Partners With Gentle Monster and Warby Parker For AI Glasses
  • www.tomsguide.com: Google's AI-powered Android XR smart glasses could be a game-changer — here's how

@zdnet.com //
Google has officially launched Flow, an AI-powered filmmaking tool designed to simplify the creation of cinematic videos. Unveiled at Google I/O 2025, Flow leverages Google's advanced AI models, including Veo for video generation, Imagen for image production, and Gemini for orchestration through natural language. This new platform is an evolution of the earlier experimental VideoFX project and aims to make it easier for storytellers to conceptualize, draft, and refine video sequences using AI. Flow provides a creative toolkit for video makers, positioning itself as a storytelling platform rather than just a simple video generator.

Flow acts as a hybrid tool that combines the strengths of Veo, Imagen, and Gemini. Veo 3, the improved video model underneath Flow, adds motion and realism meant to mimic physics, marking a step forward in dynamic content creation, even allowing for the generation of sound effects, background sounds, and character dialogue directly within videos. With Imagen, users can create visual assets from scratch and bring them into their Flow projects. Gemini helps fine-tune output, adjusting timing, mood, or even narrative arcs through conversational inputs. The platform focuses on continuity and filmmaking, allowing users to reuse characters or scenes across multiple clips while maintaining consistency.

One of Flow's major appeals is its ability to handle visual consistency, enabling scenes to blend into one another with more continuity than earlier AI systems. Filmmakers can not only edit transitions but also set camera positions, plan pans, and tweak angles. For creators frustrated by scattered generations and unstructured assets, Flow introduces a management system that organizes files, clips, and even the text used to create them. Currently, Flow is accessible to users in the U.S. subscribed to either the AI Pro or AI Ultra tiers. The Pro plan includes 100 video generations per month, while Ultra subscribers receive unlimited generations and earlier access to Veo 3, which will support built-in audio, costing $249.99 monthly.

Recommended read:
References :
  • Analytics Vidhya: Google I/O 2025: AI Mode on Google Search, Veo 3, Imagen 4, Flow, Gemini Live, and More
  • TestingCatalog: Google prepares to launch Flow, a new video editing tool, at I/O 2025
  • AI & Machine Learning: Expanding Vertex AI with the next wave of generative AI media models
  • AI News | VentureBeat: Google just leapfrogged every competitor with mind-blowing AI that can think deeper, shop smarter, and create videos with dialogue
  • www.techradar.com: Google's Veo 3 marks the end of AI video's 'silent era'
  • www.zdnet.com: Google Flow is a new AI video generator meant for filmmakers - how to try it today
  • www.techradar.com: Want to be the next Spielberg? Google’s AI-powered Flow could bring your movie ideas to life
  • the-decoder.com: Google showed off a range of new features for creators, developers, and everyday users at I/O 2025, beyond its headline announcements about search and AI models.
  • Digital Information World: At its annual I/O event, Google introduced a new AI-based application called , positioned as a creative toolkit for video makers.
  • Maginative: Google has launched Flow, a new AI-powered filmmaking tool designed to simplify cinematic clip creation and scene extension using its advanced Veo, Imagen, and Gemini models.
  • www.tomsguide.com: Google Veo 3 and Flow: The future of AI filmmaking is here — here’s how it works
  • THE DECODER: Google shows AI filmmaking tool, XR glasses and launches $250 Gemini subscription

@www.searchenginejournal.com //
References: Search Engine Journal , WhatIs ,
Google is aggressively expanding its artificial intelligence capabilities across its platforms, integrating the Gemini AI model into Search, and Android XR smart glasses. The tech giant unveiled the rollout of "AI Mode" in the U.S. Search, making it accessible to all users after initial testing in the Labs division. This move signifies a major shift in how people interact with the search engine, offering a conversational experience akin to consulting with an expert.

Google is feeding its latest AI model, Gemini 2.5, into its search algorithms, enhancing features like "AI Overviews" which are now available in over 200 countries and 40 languages and are used by 1.5 billion monthly users. In addition, Gemini 2.5 Pro introduces enhanced reasoning, through Deep Think, to give deeper and more thorough responses with AI Mode with Deep Search. Google is also testing new AI-powered features, including the ability to conduct searches through live video feeds with Search Live.

Google is also re-entering the smart glasses market with Android XR-powered spectacles featuring a hands-free camera and a voice-powered AI assistant. This project, named Astra, allows users to talk back and forth with Search about what they see in real-time with their cameras. These advancements aim to create more personalized and efficient user experiences, marking a new phase in the AI platform shift and solidifying AI's position in search.

Recommended read:
References :
  • Search Engine Journal: Google Expands AI Features in Search: What You Need to Know
  • WhatIs: Google expands Gemini model, Search as AI rivals encroach
  • www.theguardian.com: Google unveils ‘AI Mode’ in the next phase of its journey to change search

Eric Hal@techradar.com //
Google I/O 2025 saw the unveiling of 'AI Mode' for Google Search, signaling a significant shift in how the company approaches information retrieval and user experience. The new AI Mode, powered by the Gemini 2.5 model, is designed to offer more detailed results, personal context, and intelligent assistance. This upgrade aims to compete directly with the capabilities of AI chatbots like ChatGPT, providing users with a more conversational and comprehensive search experience. The rollout has commenced in the U.S. for both the browser version of Search and the Google app, although availability in other countries remains unconfirmed.

AI Mode brings several key features to the forefront, including Deep Search, Live Visual Search, and AI-powered agents. Deep Search allows users to delve into topics with unprecedented depth, running hundreds of searches simultaneously to generate expert-level, fully-cited reports in minutes. With Search Live, users can leverage their phone's camera to interact with Search in real-time, receiving context-aware responses from Gemini. Google is also bringing agentic capabilities to Search, allowing users to perform tasks like booking tickets and making reservations directly through the AI interface.

Google’s revamp of its AI search service appears to be a response to the growing popularity of AI-driven search experiences offered by companies like OpenAI and Perplexity. According to Gartner analyst Chirag Dekate, evidence suggests a greater reliance on search and AI-infused search experiences. As AI Mode rolls out, Google is encouraging website owners to optimize their content for AI-powered search by creating unique, non-commodity content and ensuring that their sites meet technical requirements and provide a good user experience.

Recommended read:
References :
  • Search Engine Journal: Google's new AI Mode in Search, integrating Gemini 2.5, aims to enhance user interaction by providing more conversational and comprehensive responses.
  • www.techradar.com: Google just got a new 'Deep Think' mode – and 6 other upgrades
  • WhatIs: Google expands Gemini model, Search as AI rivals encroach
  • www.tomsguide.com: Google Search gets an AI tab — here’s what it means for your searches
  • AI News | VentureBeat: Inside Google’s AI leap: Gemini 2.5 thinks deeper, speaks smarter and codes faster
  • Search Engine Journal: Google Gemini upgrades include Chrome integration, Live visual tools, and enhanced 2.5 models. Learn how these AI advances could reshape your marketing strategy.
  • Google DeepMind Blog: Gemini 2.5: Our most intelligent models are getting even better
  • learn.aisingapore.org: Updates to Gemini 2.5 from Google DeepMind
  • THE DECODER: Google upgrades Gemini 2.5 Pro with a new Deep Think mode for advanced reasoning abilities
  • www.techradar.com: I've been using Google's new AI mode for Search – here's how to master it
  • www.theguardian.com: Search engine revamp and Gemini 2.5 introduced at conference in latest showing tech giant is all in on AI on Tuesday unleashed another wave of technology to accelerate a year-long makeover of its search engine that is changing the way people get information and curtailing the flow of internet traffic to other websites.
  • LearnAI: Updates to Gemini 2.5 from Google DeepMind
  • www.analyticsvidhya.com: Google I/O 2025: AI Mode on Google Search, Veo 3, Imagen 4, Flow, Gemini Live, and More
  • techvro.com: Google AI Mode Promises Deep Search and Goes Beyond AI Overviews
  • THE DECODER: Google pushes AI-powered search with agents, multimodality, and virtual shopping
  • felloai.com: Google I/O 2025 Recap With All The Jaw-Dropping AI Announcements
  • Analytics Vidhya: Google I/O 2025: AI Mode on Google Search, Veo 3, Imagen 4, Flow, Gemini Live, and More
  • LearnAI: Gemini as a universal AI assistant
  • Fello AI: Google I/O 2025 Recap With All The Jaw-Dropping AI Announcements
  • AI & Machine Learning: Today at Google I/O, we're expanding that help enterprises build more sophisticated and secure AI-driven applications and agents
  • www.techradar.com: Google Gemini 2.5 Flash promises to be your favorite AI chatbot, but how does it compare to ChatGPT 4o?
  • www.laptopmag.com: From $250 AI subscriptions to futuristic glasses and search that talks back, here’s what people are saying about Tuesday's Google I/O.
  • www.tomsguide.com: Google’s Gemini AI can now access Gmail, Docs, Drive, and more to deliver personalized help — but it raises new privacy concerns.
  • Data Phoenix: Google updated its model lineup and introduced a 'Deep Think' reasoning mode for Gemini 2.5 Pro
  • Maginative: Google’s revamped Canvas, powered by the Gemini 2.5 Pro model, lets you turn ideas into apps, quizzes, podcasts, and visuals in seconds—no code required.
  • Tech News | Euronews RSS: The tech giant is introducing a new "AI mode" that will embed chatbot capabilities into its search engine to keep up with rivals like OpenAI's ChatGPT.
  • learn.aisingapore.org: Advancing Gemini’s security safeguards – Google DeepMind
  • Data Phoenix: Google has launched major Gemini updates, including free visual assistance via Gemini Live, new subscription tiers starting at $19.99/month, advanced creative tools like Veo 3 for video generation with native audio, and an upcoming autonomous Agent Mode for complex task management.
  • www.zdnet.com: Everything from Google I/O 2025 you might've missed: Gemini, smart glasses, and more
  • thetechbasic.com: Google now adds ads to AI Mode and AI Overviews in search
  • Google DeepMind Blog: Gemini 2.5: Our most intelligent models are getting even better

Alexey Shabanov@TestingCatalog //
Google has launched the NotebookLM mobile app for Android and iOS, bringing its AI-powered research assistant to mobile devices. This release marks a significant step in expanding access to NotebookLM, which was initially launched as a web-based tool in 2023 under the codename "Project Tailwind." The mobile app aims to offer personalized learning and efficient content synthesis, allowing users to interact with and process information on the go. The app is officially available to everyone after months of waiting, offering the core features of NotebookLM, with the promise of continued functionality additions.

The NotebookLM mobile app focuses on audio-first experiences, with features like audio overviews that generate podcast-style summaries. These summaries can be played directly from the list view without opening a project, making it feel like a media player for casual content consumption. Users can also download audio overviews for offline playback and listen in the background, supporting learning during commutes or other activities. Moreover, the app supports interactive mode in audio sessions, where users can ask questions mid-playback, creating a live dialogue experience.

The mobile app retains the functionality of the web version, including the ability to create new notebooks and upload sources like PDFs, Google Docs, and YouTube videos. Users can add sources directly from their mobile devices by using the "Share" button in any app, making it easier to build and maintain research libraries. NotebookLM relies only on user-uploaded sources, ensuring reliable and verifiable information. The rollout underscores Google’s evolving strategy for NotebookLM, transitioning from a productivity assistant to a multimodal content platform, appealing to students, researchers, and content creators seeking flexible ways to absorb structured knowledge.

Recommended read:
References :
  • AI News | VentureBeat: Google finally launches NotebookLM mobile app at I/O: hands-on, first impressions
  • www.laptopmag.com: An exclusive look at Google's NotebookLM app on Android and iOS
  • TestingCatalog: Google launches NotebookLM mobile app with audio-first features on mobile
  • www.tomsguide.com: NotebookLM just arrived on Android — and it can turn your notes into podcasts
  • THE DECODER: Google launches NotebookLM mobile app for Android and iOS
  • MarkTechPost: Google AI Releases Standalone NotebookLM Mobile App with Offline Audio and Seamless Source Integration
  • the-decoder.com: Google launches NotebookLM mobile app for Android and iOS
  • www.marktechpost.com: Google AI Releases Standalone NotebookLM Mobile App with Offline Audio and Seamless Source Integration
  • www.techradar.com: Google's free NotebookLM AI app is out now for Android and iOS – here's why it's a day-one download for me

Alexey Shabanov@TestingCatalog //
Google has officially launched the NotebookLM mobile app for both Android and iOS, extending the reach of its AI-powered research assistant. This release, anticipated before Google I/O 2025, allows users to leverage NotebookLM's capabilities directly from their smartphones and tablets. The app aims to help users understand information more effectively, regardless of their location, marking a step towards broader accessibility to AI tools.

The NotebookLM mobile app provides a range of features, including the ability to create new notebooks and add various content types, such as PDFs, websites, YouTube videos, and text. A key feature highlighted by Google is the availability of "Audio Overviews," which automatically generates audio summaries for offline and background playback. Furthermore, users can interact with AI hosts (in beta) to ask follow-up questions, enhancing the learning and research experience on the go. The app also integrates with the Android and iOS share sheets for quickly adding sources.

The initial release offers a straightforward user interface optimized for both phones and tablets. Navigation within the app includes a bottom bar providing easy access to Sources, Chat Q&A, and Studio. While it currently doesn't fully utilize Material 3 design principles, Google emphasizes this is an early version. Users can now download the NotebookLM app from the Google Play Store and the App Store, fulfilling a top feature request. Google has indicated that additional updates and features are planned for future releases.

Recommended read:
References :
  • TestingCatalog: Google launches NotebookLM mobile app with audio-first features on mobile
  • The Official Google Blog: Understand anything, anywhere with the new NotebookLM app
  • www.laptopmag.com: An exclusive look at Google's NotebookLM app on Android and iOS
  • www.tomsguide.com: NotebookLM just arrived on Android — and it can turn your notes into podcasts
  • www.marktechpost.com: Google AI Releases Standalone NotebookLM Mobile App with Offline Audio and Seamless Source Integration.
  • THE DECODER: Google launches NotebookLM mobile app with audio-first features on mobile
  • 9to5Mac: Google launches NotebookLM mobile app for Android and iOS
  • TechCrunch: Google launches stand-alone NotebookLM app for Android
  • AI News | VentureBeat: Google finally launches NotebookLM mobile app at I/O: hands-on, first impressions
  • MarkTechPost: Google AI Releases Standalone NotebookLM Mobile App with Offline Audio and Seamless Source Integration
  • Dataconomy: Google brings NotebookLM to mobile with new standalone apps
  • The Tech Basic: Google launches NotebookLM apps letting users research on the go
  • the-decoder.com: Google launches NotebookLM mobile app for Android and iOS
  • www.techradar.com: Google's free NotebookLM AI app is out now for Android and iOS – here's why it's a day-one download for me
  • thetechbasic.com: Google Launches NotebookLM Apps Letting Users Research on the Go
  • www.techradar.com: Google I/O 2025 live blog covering all the major announcements, including Android XR updates, new Gemini features, and more.