News from the AI & ML world

DeeperML - #ondeviceai

Alexey Shabanov@TestingCatalog //
Google is aggressively integrating its Gemini AI model across a multitude of platforms, signaling a significant push towards embedding AI into everyday technologies. The initiatives span from enhancing user experiences in applications like Google Photos to enabling advanced capabilities in robotics and providing developers with powerful coding tools via the Gemini CLI. This widespread integration highlights Google's vision for a future where AI is a seamless and integral part of various technological ecosystems.

The integration of Gemini into Google Photos is designed to improve search functionality, allowing users to find specific images more efficiently using natural language queries. Similarly, the development of on-device Gemini models for robotics addresses critical concerns around privacy and latency, ensuring that robots can operate effectively even without a constant internet connection. This is particularly crucial for tasks requiring real-time decision-making, where delays could pose significant risks.

Furthermore, Google's release of the Gemini CLI provides developers with an open-source AI agent directly accessible from their terminal. This tool supports various coding and debugging tasks, streamlining the development process. Additionally, Gemini models are being optimized for edge deployment, allowing for AI functionality in environments with limited or no cloud connectivity, further demonstrating Google's commitment to making AI accessible and versatile across diverse applications.

Recommended read:
References :
  • www.tomsguide.com: Google's 'Ask Photos' AI search is back and should be better than ever.
  • www.techradar.com: Google’s new Gemini AI model means your future robot butler will still work even without Wi‑Fi.
  • Maginative: Google Announces On-Device Gemini Robotics Model
  • www.marktechpost.com: Google AI Releases Gemini CLI: An Open-Source AI Agent for Your Terminal
  • TestingCatalog: Google prepares interactive Storybook experience for Gemini users
  • felloai.com: Information on Google’s Gemini 3.0 and what to expect from the new model.
  • www.marktechpost.com: Getting started with Gemini Command Line Interface (CLI)
  • Maginative: Google Launches Gemini CLI, an open source AI Agent in your terminal

Mark Gurman@Bloomberg Technology //
Apple is facing delays in the release of its AI-powered Siri upgrade, now reportedly slated for Spring 2026 with the iOS 26.4 update. This news follows the recent WWDC 2025 event, where AI features were showcased across various Apple operating systems, but the highly anticipated Siri overhaul was notably absent. Sources indicate that the delay stems from challenges in integrating older Siri systems with newer platforms, forcing engineers to rebuild the assistant from scratch. Craig Federighi, Apple’s head of software engineering, explained that the previous V1 architecture was insufficient for achieving the desired quality, prompting a shift to a "deeper end-to-end architecture" known as V2.

This delay has also reportedly caused internal tensions within Apple, with the AI and marketing teams allegedly blaming each other for overpromising and failing to meet timelines. While no exact date has been finalized for the iOS 26.4 release, insiders suggest a spring timeframe, aligning with Apple's typical release schedule for ".4" updates. The upgraded Siri is expected to offer smarter responses, improved app control, and on-screen awareness, allowing it to tap into users' personal context and perform actions based on what's displayed on their devices.

Separately, Apple researchers have revealed structural failures in large reasoning models (LRMs) through puzzle-based evaluations. A recently released Apple research paper claimed that contemporary AI LLMs and LRMs fail to make sound judgements as the complexity of problems in controlled puzzle environments they were tasked to solve increased, revealing their fundamental limitations and debunking the common belief that these models can think like a human being. This work, conducted using puzzles like the Tower of Hanoi and River Crossing, aimed to assess the true reasoning capabilities of AI models by analyzing their performance on unfamiliar tasks, free from data contamination. Professor Seok Joon Kwon of Sungkyunkwan University believes Apple does not have enough high-performance hardware to test what high-end LRMs and LLMs are truly capable of.

Recommended read:
References :
  • Bloomberg Technology: Apple targets spring 2026 for release of delayed Siri AI upgrade
  • PCMag Middle East ai: Apple Explains Why It Delayed AI Siri, Confirms It Won't Arrive Until 2026
  • www.marktechpost.com: Apple Researchers Reveal Structural Failures in Large Reasoning Models Using Puzzle-Based Evaluation
  • www.techradar.com: Apple reportedly targets 'spring 2026' for launch of delayed AI Siri upgrade – but is that too late?
  • www.tomsguide.com: Siri may not get an AI upgrade until next Spring — what we know
  • thetechbasic.com: Apple AI Roadmap: Contextual Siri 2026, Knowledge Chatbot & Copilot
  • PCMag Middle East ai: Report: Siri's Long-Delayed AI Features May Arrive With iOS 26.4
  • AppleMagazine: Apple Targets iOS 26.4 for Siri AI Upgrade in March 2026
  • Mark Gurman: NEW: Apple has set an internal release target of spring 2026 for its delayed upgrade of Siri, marking a key step in its artificial intelligence turnaround effort. Plus, the latest on the company's efforts here.
  • The Tech Basic: Apple AI Roadmap: Contextual Siri 2026, Knowledge Chatbot & Copilot
  • www.eweek.com: ‘This Work Needed More Time’: Apple Delays Siri Upgrade to Spring 2026
  • eWEEK: ‘This Work Needed More Time’: Apple Delays Siri Upgrade to Spring 2026
  • Analytics India Magazine: Apple sets target to release delayed Siri AI update by Spring 2026
  • www.laptopmag.com: Apple’s AI-powered Siri reportedly has a new target date. Will it stick this time?

Ashutosh Singh@The Tech Portal //
Google has launched AI Edge Gallery, an open-source platform aimed at developers who want to deploy AI models directly on Android devices. This new platform allows for on-device AI execution using tools like LiteRT and MediaPipe, supporting models from Hugging Face. With future support for iOS planned, AI Edge Gallery emphasizes data privacy and low latency by eliminating the need for cloud connectivity, making it ideal for industries that require local processing of sensitive data.

The AI Edge Gallery app, released under the Apache 2.0 license and hosted on GitHub, is currently in an experimental Alpha release. The app integrates Gemma 3 1B, a compact 529MB language model, capable of processing up to 2,585 tokens per second on mobile GPUs, enabling tasks like text generation and image analysis in under a second. By using Google’s AI Edge platform, developers can leverage tools like MediaPipe and TensorFlow Lite to optimize model performance on mobile devices. The company is actively seeking feedback from developers and users.

AI Edge Gallery contains categories like ‘AI Chat’ and ‘Ask Image’ to guide users to relevant tools, as well as a ‘Prompt Lab’ for testing and refining prompts. On-device AI processing ensures that complex AI tasks can be performed without transmitting data to external servers, reducing potential security risks and improving response times. While newer devices with high-performance chips can run models smoothly, older phones may experience lag. Google is also planning to launch the app on iOS soon.

Recommended read:
References :
  • The Tech Portal: Google rolls out ‘AI Edge Gallery’ app for Android that lets you run AI models locally on device
  • www.infoworld.com: Google’s AI Edge Gallery will let developers deploy offline AI models — here’s how it works
  • Latest news: This new Google app lets you use AI on your phone without the internet - here's how
  • developers.googleblog.com: The 529MB Gemma 3 1B model delivers up to 2,585 tokens per second during on mobile GPUs, enabling sub-second tasks like text generation and image analysis.
  • venturebeat.com: Google quietly launched AI Edge Gallery, an experimental Android app that runs AI models offline without internet, bringing Hugging Face models directly to smartphones with enhanced privacy.

Ashutosh Singh@The Tech Portal //
Google has launched the 'AI Edge Gallery' app for Android, with plans to extend it to iOS soon. This innovative app enables users to run a variety of AI models locally on their devices, eliminating the need for an internet connection. The AI Edge Gallery integrates models from Hugging Face, a popular AI repository, allowing for on-device execution. This approach not only enhances privacy by keeping data on the device but also offers faster processing speeds and offline functionality, which is particularly useful in areas with limited connectivity.

The app uses Google’s AI Edge platform, which includes tools like MediaPipe and TensorFlow Lite, to optimize model performance on mobile devices. A key model utilized is Gemma 31B, a compact language model designed for mobile platforms that can process data rapidly. The AI Edge Gallery features an interface with categories like ‘AI Chat’ and ‘Ask Image’ to help users find the right tools. Additionally, a ‘Prompt Lab’ is available for users to experiment with and refine prompts.

Google is emphasizing that the AI Edge Gallery is currently an experimental Alpha release and is encouraging user feedback. The app is open-source under the Apache 2.0 license, allowing for free use, including for commercial purposes. However, the performance of the app may vary based on the device's hardware capabilities. While newer phones with advanced processors can run models smoothly, older devices might experience lag, particularly with larger models.

In related news, Google Cloud has introduced advancements to BigLake, its storage engine designed to create open data lakehouses on Google Cloud that are compatible with Apache Iceberg. These enhancements aim to eliminate the need to sacrifice open-format flexibility for high-performance, enterprise-grade storage management. These updates include Open interoperability across analytical and transactional systems:

The BigLake metastore provides the foundation for interoperability, allowing you to access all your Cloud Storage and BigQuery storage data across multiple runtimes including BigQuery, AlloyDB (preview), and open-source, Iceberg-compatible engines such as Spark and Flink.New, high-performance Iceberg-native Cloud Storage:

We are simplifying lakehouse management with automatic table maintenance (including compaction and garbage collection) and integration with Google Cloud Storage management tools, including auto-class tiering and encryption.

Recommended read:
References :
  • Data Analytics: BigLake evolved: Build open, high-performance, enterprise Iceberg-native lakehouses
  • The Tech Portal: Google rolls out ‘AI Edge Gallery’ app for Android that lets you run AI models locally on device
  • www.infoworld.com: Google Cloud’s BigLake-driven lakehouse updates aim to optimize performance, costs
  • TechCrunch: Last week, Google quietly released an app that lets users run a range of openly available AI models from the AI dev platform Hugging Face on their phones.
  • Neowin: Google's new AI Edge Gallery app brings offline AI to your Android (and soon iOS) device
  • www.infoworld.com: Google has launched AI Edge Gallery, an open-source platform that enables developers to run advanced AI models directly on Android devices, with iOS support planned for a future release.
  • Latest news: This new Google app lets you use AI on your phone without the internet - here's how
  • Techzine Global: New Google app runs AI offline on smartphones
  • venturebeat.com: Google quietly launches AI Edge Gallery, letting Android phones run AI without the cloud
  • Dataconomy: Google released the Google AI Edge Gallery app last week, enabling users to download and run AI models from Hugging Face on their phones.