News from the AI & ML world

DeeperML - #robots

Michael Nuñez@venturebeat.com //
Nvidia has made significant strides in enhancing robot training and AI capabilities, unveiling innovative solutions at its GTC conference. A key announcement was Cosmos-Transfer1, a groundbreaking AI model designed to generate photorealistic simulations for training robots and autonomous vehicles. This model bridges the gap between virtual training environments and real-world applications by using multimodal inputs to create highly realistic simulations. This adaptive multimodal control system allows developers to weight different visual inputs, such as depth or object boundaries, to improve the realism and utility of the generated environments.

Nvidia also introduced its next-generation GPU superchips, including the second generation of the Grace Blackwell chip and the Vera Rubin, expected in the second half of 2026. The Vera Rubin will feature 288GB of high-bandwidth memory 4 (HBM4) and will be paired with CPUs boasting 88 custom Arm cores. These new chips promise substantial increases in compute capacity, with Rubin delivering a 900x speedup compared to the previous generation Hopper chips. This positions Nvidia to tackle the increasing demands of generative AI workloads, including training massive AI models and running inference workloads.

Share: bluesky twitterx--v2 facebook--v1 threads


References :
  • venturebeat.com: Nvidia’s Cosmos-Transfer1 makes robot training freakishly realistic—and that changes everything
  • AIwire: Nvidia Touts Next Generation GPU Superchip and New Photonic Switches
  • www.laptopmag.com: Blackwell Ultra and Rubin Ultra are Nvidia's newest additions to the growing list of AI superchips
Classification:
  • HashTags: #Nvidia #Robots #CosmosTransfer1
  • Company: Nvidia
  • Target: Robotics
  • Product: Cosmos-Transfer1
  • Feature: Robot training
  • Type: AI
  • Severity: Informative
@Google DeepMind Blog //
Google is pushing the boundaries of AI and robotics with its Gemini AI models. Gemini Robotics, an advanced vision-language-action model, now enables robots to perform physical tasks with improved generalization, adaptability, and dexterity. This model interprets and acts on text, voice, and image data, showcasing Google's advancements in integrating AI for practical applications. Furthermore, the development of Gemini Robotics-ER, which incorporates embodied reasoning capabilities, signifies another step toward smarter, more adaptable robots.

Google's approach to robotics emphasizes safety, employing both physical and semantic safety systems. The company is inviting filmmakers and creators to experiment with the model to improve the design and development. Veo builds on years of generative video model work, including Generative Query Network(GQN),DVD-GAN,Imagen-Video,Phenaki,WALT,VideoPoetandLumiere— combining architecture, scaling laws and other novel techniques to improve quality and output resolution.

Share: bluesky twitterx--v2 facebook--v1 threads


References :
  • Google DeepMind Blog: Gemini Robotics brings AI into the physical world
  • Maginative: Google DeepMind Unveils Gemini Robotics Models to Bridge AI and Physical World
  • IEEE Spectrum: With Gemini Robotics, Google Aims for Smarter Robots
  • The Official Google Blog: Take a closer look at our new Gemini models for robotics.
  • THE DECODER: Google Deepmind unveils new AI models for robotic control
  • www.tomsguide.com: Google is putting it's Gemini 2.0 AI into robots — here's how it's going
  • Verdict: Google DeepMind unveils Gemini AI models for robotics
  • MarkTechPost: Google DeepMind’s Gemini Robotics: Unleashing Embodied AI with Zero-Shot Control and Enhanced Spatial Reasoning
  • LearnAI: Research Published 12 March 2025 Authors Carolina Parada Introducing Gemini Robotics, our Gemini 2.0-based model designed for robotics At Google DeepMind, we’ve been making progress in how our Gemini models solve complex problems through multimodal reasoning across text, images, audio and video. So far however, those abilities have been largely confined to the digital realm....
  • OODAloop: Google DeepMind unveils new AI models for robotic control.
  • www.producthunt.com: Gemini Robotics
  • Last Week in AI: Last Week in AI #303 - Gemini Robotics, Gemma 3, CSM-1B
  • Windows Copilot News: Google is prepping Gemini to take action inside of apps
  • Last Week in AI: Discusses Gemini Robotics in the context of general AI agents and robotics.
  • www.infoq.com: Google DeepMind unveils Gemini Robotics, an advanced AI model for enhancing robotics through vision, language, and action.
  • AI & Machine Learning: This article discusses how generative AI is poised to revolutionize multiplayer games, offering personalized experiences through dynamic narratives and environments. The article specifically mentions Google's Gemini AI model and its potential to enhance gameplay.
  • Gradient Flow: This podcast episode discusses various advancements in AI, including Google's Gemini Robotics and Gemma 3, as well as the evolving regulatory landscape across different countries.
  • Insight Partners: This article highlights Andrew Ng's keynote at ScaleUp:AI '24, where he discusses the exciting trends in AI agents and applications, mentioning Google's Gemini AI assistant and its role in driving innovation.
  • www.tomsguide.com: You can now use Google Gemini without an account — here's how to get started
Classification:
@www.it-daily.net //
Apple is reportedly delving into the realm of robotics, exploring both humanoid and non-humanoid robot concepts. According to analyst Ming-Chi Kuo, the tech giant is in the early stages of research and development in this area, though the company itself has not officially confirmed these plans. A recently published research paper by Apple on human-robot interaction, specifically using a Pixar-like desk lamp as an example, supports these findings.

Apple's foray into robotics signifies a diversification of its technological portfolio. While still in the early stages, with mass production potentially not expected until 2028 at the earliest, the company seems to be focusing on the human-robot interaction and the underlying sensor and software technology. The smart home ecosystem is a potential area of application, with possibilities ranging from intelligent assistance systems to more complex humanoid robots.

Share: bluesky twitterx--v2 facebook--v1 threads


References :
Classification: