Sam Khosroshahi@lambdalabs.com
//
References:
Fello AI
, lambdalabs.com
,
NVIDIA is pushing the boundaries of artificial intelligence in healthcare and robotics, introducing several groundbreaking advancements. One notable innovation is the DNA LLM, designed to decode the complex genetic information found in DNA, RNA, and proteins. This tool aims to transform genomic research, potentially leading to new understandings and treatments for various diseases.
The company's commitment to AI extends to robotics with the release of Isaac GR00T N1, an open-source platform for humanoid robots. This initiative is expected to accelerate innovation in the field, providing developers with the resources needed to create more advanced and capable robots. Additionally, an NVIDIA research team has developed Hymba, a family of small language models that combine transformer attention with state space models, surpassing the Llama-3.2-3B model in performance while significantly reducing cache size and increasing throughput. Recommended read:
References :
Dean Takahashi@AI News | VentureBeat
//
NVIDIA, Google DeepMind, and Disney Research are collaborating on Newton, an open-source physics engine designed to advance robot learning, enhance simulation accuracy, and facilitate the development of next-generation robotic characters. Newton is built on NVIDIA’s Warp framework and aims to provide a scalable, high-performance simulation environment optimized for AI-driven humanoid robots. MuJoCo-Warp, a collaboration with Google DeepMind, accelerates robotics workloads by over 70x, while Disney plans to integrate Newton into its robotic character platform for expressive, interactive robots.
The engine's creation is intended to bridge the gap between simulation and real-world robotics. NVIDIA will also supercharge humanoid robot development with the Isaac GR00T N1 foundation model for human-like reasoning. Newton is built on NVIDIA Warp, a CUDA-based acceleration library that enables GPU-powered physics simulations. Newton is also optimized for robot learning frameworks, including MuJoCo Playground and NVIDIA Isaac Lab, making it an essential tool for developers working on generalist humanoid robots. This initiative is part of NVIDIA's broader effort to accelerate physical AI progress. Recommended read:
References :
Michael Nuñez@venturebeat.com
//
References:
venturebeat.com
, AIwire
,
Nvidia has made significant strides in enhancing robot training and AI capabilities, unveiling innovative solutions at its GTC conference. A key announcement was Cosmos-Transfer1, a groundbreaking AI model designed to generate photorealistic simulations for training robots and autonomous vehicles. This model bridges the gap between virtual training environments and real-world applications by using multimodal inputs to create highly realistic simulations. This adaptive multimodal control system allows developers to weight different visual inputs, such as depth or object boundaries, to improve the realism and utility of the generated environments.
Nvidia also introduced its next-generation GPU superchips, including the second generation of the Grace Blackwell chip and the Vera Rubin, expected in the second half of 2026. The Vera Rubin will feature 288GB of high-bandwidth memory 4 (HBM4) and will be paired with CPUs boasting 88 custom Arm cores. These new chips promise substantial increases in compute capacity, with Rubin delivering a 900x speedup compared to the previous generation Hopper chips. This positions Nvidia to tackle the increasing demands of generative AI workloads, including training massive AI models and running inference workloads. Recommended read:
References :
Shelly Palmer@Shelly Palmer
//
References:
Shelly Palmer
, Upward Dynamism
AI agents are rapidly evolving, becoming a focal point in the tech industry. OpenAI has released an API designed to make the deployment of AI agents significantly easier, opening up new possibilities for developers. These agents are capable of tasks like searching the web in real-time and analyzing extensive datasets, heralding a new era of automation and productivity. Experts predict substantial growth in the AI agent market, forecasting an increase from $5 billion in 2024 to over $47 billion by 2030.
AI agents are being designed as smart software programs that can understand and act independently to achieve assigned goals. Unlike chatbots or rigid scripts, AI agents break down complex tasks into manageable steps, adapt their strategies as needed, and learn from their experiences. Core elements include tasks, models (like GPT or Claude), memory for storing context, and external tools such as APIs and web access. These components work together iteratively, planning, acting, and adjusting as necessary without constant human intervention. Recommended read:
References :
@Google DeepMind Blog
//
Google is pushing the boundaries of AI and robotics with its Gemini AI models. Gemini Robotics, an advanced vision-language-action model, now enables robots to perform physical tasks with improved generalization, adaptability, and dexterity. This model interprets and acts on text, voice, and image data, showcasing Google's advancements in integrating AI for practical applications. Furthermore, the development of Gemini Robotics-ER, which incorporates embodied reasoning capabilities, signifies another step toward smarter, more adaptable robots.
Google's approach to robotics emphasizes safety, employing both physical and semantic safety systems. The company is inviting filmmakers and creators to experiment with the model to improve the design and development. Veo builds on years of generative video model work, including Generative Query Network(GQN),DVD-GAN,Imagen-Video,Phenaki,WALT,VideoPoetandLumiere— combining architecture, scaling laws and other novel techniques to improve quality and output resolution. Recommended read:
References :
Tris Warkentin@The Official Google Blog
//
Google AI has released Gemma 3, a new family of open-source AI models designed for efficient and on-device AI applications. Gemma 3 models are built with technology similar to Gemini 2.0, intended to run efficiently on a single GPU or TPU. The models are available in various sizes: 1B, 4B, 12B, and 27B parameters, with options for both pre-trained and instruction-tuned variants, allowing users to select the model that best fits their hardware and specific application needs.
Gemma 3 offers practical advantages including efficiency and portability. For example, the 27B version has demonstrated robust performance in evaluations while still being capable of running on a single GPU. The 4B, 12B, and 27B models are capable of processing both text and images, and supports more than 140 languages. The models have a context window of 128,000 tokens, making them well suited for tasks that require processing large amounts of information. Google has built safety protocols into Gemma 3, including a safety checker for images called ShieldGemma 2. Recommended read:
References :
ChinaTechNews.com Staff@ChinaTechNews.com
//
Clone Robotics, a Polish startup, has unveiled Protoclone, a human-like robot designed with synthetic muscles and a skeletal structure mimicking human anatomy. The Protoclone is described as a faceless android with over 200 degrees of freedom, more than 1,000 artificial muscle fibers (Myofibers), and 500 sensors. According to the company, Protoclone operates with systems that closely replicate human muscular, skeletal, vascular, and nervous systems, even featuring 206 bones made from polymers.
Reactions to the Protoclone have been mixed, with some social media users expressing unease over its lifelike movements and appearance. Clone Robotics plans to make the android available for preorder later this year. The company envisions the robot could perform household chores in the future. Protoclone uses a pneumatic system to control its artificial muscles but the final product will use hydraulics. The robot uses a system of sensors linked to four cameras to navigate. Recommended read:
References :
@www.therobotreport.com
//
Figure AI has unveiled Helix, a new AI system designed to power humanoid robots with advanced movement capabilities. Helix aims to enable robots to perform complex tasks and handle unfamiliar objects through voice commands. The system combines a 7-billion-parameter multimodal language model, which acts as the robot's "brain" by processing speech and visual data, with an 80-million-parameter AI that translates these instructions into precise movements.
Helix can control 35 degrees of freedom, from individual finger movements to torso control. Demonstrations have showcased robots responding to voice commands, identifying objects, and grasping them accurately, even collaborating to place food items into a refrigerator without specific prior training. The system required only 500 hours of training data and runs on embedded GPUs within the robots. Figure AI is also reportedly seeking $1.5 billion in funding, potentially valuing the company at $39.5 billion. This comes after the company ended its collaboration with OpenAI on robot-specific AI models, though OpenAI remains a significant investor. Figure CEO Brett Adcock believes Helix is crucial for scaling robots in household settings, allowing them to adapt to new situations without requiring constant reprogramming. Recommended read:
References :
@www.reuters.com
//
References:
Engineering at Meta
, www.eweek.com
,
Meta is expanding its artificial intelligence research into the realm of humanoid robotics, aiming to develop AI-driven software and sensors. This initiative focuses on creating intelligent machines that can interact with the physical world, potentially powering consumer robots. The company's efforts are concentrated on "embodied AI," which combines intelligence with real-world interactions, enabling robots to move, sense, and make decisions in three-dimensional environments.
Meta is not initially planning to release its own branded robots. Instead, the company is concentrating on developing AI-powered software and sensor technology that can be utilized by other robotics manufacturers. This strategy positions Meta alongside tech giants like Tesla, Apple, and Google, all of which are also investing in the robotics sector. Meta is also prioritizing user data protection by using source code analysis to detect and prevent unauthorized data scraping across its platforms, including Facebook, Instagram, and Reality Labs. Recommended read:
References :
@oodaloop.com
//
References:
OODAloop
, TechCrunch
,
Symbotic, a robotics firm, has announced it is acquiring Walmart’s automation division for an initial payment of $200 million, with the potential for an additional $350 million based on the deal's performance. This acquisition solidifies Symbotic's role as a key technology supplier to Walmart and significantly expands its presence in the automation market. The agreement sees Walmart purchasing automation systems from Symbotic, which avoids the prospect of Walmart developing rival automation solutions in-house. The partnership is not just an aquisition, it is expected to provide Symbotic with immediate revenue and strengthens an already established relationship with the retail giant.
The deal will see Symbotic take control of automating Walmart’s pickup and delivery centers, with Walmart funding a development program worth $520 million to support the technology. This includes the initial $200m, with the remainder in future payments. Walmart aims to improve customer experience through quicker, more efficient service with the implementation of Symbotic's automated systems. This strategic move builds on a relationship that began in 2017 and positions Symbotic to enhance Walmart's in-store Accelerated Pickup and Delivery capabilities. The transaction is projected to close in the second quarter of 2025. Recommended read:
References :
@www.digitimes.com
//
References:
blogs.nvidia.com
, analyticsindiamag.com
,
Nvidia is significantly advancing its role in humanoid robotics through its Jetson Thor platform, aiming to replicate its AI chip success. CEO Jensen Huang has identified humanoid robots as a key area for large-scale production. To further this goal, Nvidia is releasing Isaac GR00T Blueprint which helps with generating training data for humanoid robots. This blueprint leverages workflows for synthetic data creation and simulation, and includes tools like the GR00T-Teleop, GR00T-Mimic and GR00T-Gen to generate large training datasets for robots to learn imitation skills. The company is also releasing a collection of robot foundation models and data pipelines to accelerate development.
Foxconn is partnering with Nvidia to develop humanoid robot services, integrating Nvidia's advanced software and hardware in Kaohsiung City, Taiwan. This partnership marks a significant move for Foxconn to diversify beyond contract manufacturing in electronics, with plans to collaborate on smart city projects. Additionally, Taiwan is building a supercomputer based on NVIDIA’s Blackwell architecture, further emphasizing the island's commitment to AI advancement and robotic capabilities. NVIDIA's Jetson Thor computing system, debuting in early 2025, will be central to this endeavor. The company's actions highlight a broader trend towards embodied AI. Recommended read:
References :
@analyticsindiamag.com
//
References:
www.digitimes.com
, analyticsindiamag.com
,
Nvidia is aggressively expanding its presence in the robotics sector, highlighted by the development of its Jetson Thor computing platform. This platform is designed to replicate Nvidia’s success in AI chips and is a core component of their strategy to develop advanced humanoid robots. Furthermore, Nvidia is not working alone in this endeavor. They have partnered with Foxconn to create humanoid robots, aiming to move beyond just manufacturing and integrate into new tech areas. This strategic move demonstrates Nvidia’s focus on becoming a dominant player in AI-driven robotics, specifically for humanoid technology.
Nvidia is also addressing the challenge of training these robots through their Isaac GR00T Blueprint, unveiled at CES. This blueprint utilizes synthetic data generation to create the extensive datasets needed for imitation learning, allowing robots to mimic human actions. A new workflow uses Apple Vision Pro to capture human actions in a digital twin and the data is used in the Isaac Lab framework, which teaches robots to move and interact safely. Nvidia’s Cosmos platform also is in use by generating physics-aware videos that are also used to train robots. The company's CEO, Jensen Huang, emphasizes humanoid robots as the next big leap in AI innovation, aiming to establish Nvidia as a key player in the future of robotics and autonomous systems. Recommended read:
References :
|
BenchmarksBlogsResearch Tools |