Home OpenAI Google DeepMind Releases Gemini Robotics On-Device: Local AI Model for Real-Time Robotic Dexterity
OpenAI

Google DeepMind Releases Gemini Robotics On-Device: Local AI Model for Real-Time Robotic Dexterity

Share
Google DeepMind Releases Gemini Robotics On-Device: Local AI Model for Real-Time Robotic Dexterity
Share


Google DeepMind has unveiled Gemini Robotics On-Device, a compact, local version of its powerful vision-language-action (VLA) model, bringing advanced robotic intelligence directly onto devices. This marks a key step forward in the field of embodied AI by eliminating the need for continuous cloud connectivity while maintaining the flexibility, generality, and high precision associated with the Gemini model family.

Local AI for Real-World Robotic Dexterity

Traditionally, high-capacity VLA models have relied on cloud-based processing due to computational and memory constraints. With Gemini Robotics On-Device, DeepMind introduces an architecture that operates entirely on local GPUs embedded within robots, supporting latency-sensitive and bandwidth-constrained scenarios like homes, hospitals, and manufacturing floors.

The on-device model retains the core strengths of Gemini Robotics: the ability to understand human instructions, perceive multimodal input (visual and textual), and generate real-time motor actions. It is also highly sample-efficient, requiring only 50 to 100 demonstrations to generalize new skills, making it practical for real-world deployment across varied settings.

Core Features of Gemini Robotics On-Device

  1. Fully Local Execution: The model runs directly on the robot’s onboard GPU, enabling closed-loop control without internet dependency.
  2. Two-Handed Dexterity: It can execute complex, coordinated bimanual manipulation tasks, thanks to its pretraining on the ALOHA dataset and subsequent finetuning.
  3. Multi-Embodiment Compatibility: Despite being trained on specific robots, the model generalizes across different platforms including humanoids and industrial dual-arm manipulators.
  4. Few-Shot Adaptation: The model supports rapid learning of novel tasks from a handful of demonstrations, dramatically reducing development time.

Real-World Capabilities and Applications

Dexterous manipulation tasks such as folding clothes, assembling components, or opening jars demand fine-grained motor control and real-time feedback integration. Gemini Robotics On-Device enables these capabilities while reducing communication lag and improving responsiveness. This is particularly critical for edge deployments where connectivity is unreliable or data privacy is a concern.

Potential applications include:

  • Home assistance robots capable of performing daily chores.
  • Healthcare robots that assist in rehabilitation or eldercare.
  • Industrial automation systems requiring adaptive assembly line workers.

SDK and MuJoCo Integration for Developers

Alongside the model, DeepMind has released a Gemini Robotics SDK that provides tools for testing, fine-tuning, and integrating the on-device model into custom workflows. The SDK supports:

  • Training pipelines for task-specific tuning.
  • Compatibility with various robot types and camera setups.
  • Evaluation within the MuJoCo physics simulator, which has been open-sourced with new benchmarks specifically designed for assessing bimanual dexterity tasks.

The combination of local inference, developer tools, and robust simulation environments positions Gemini Robotics On-Device as a modular, extensible solution for robotics researchers and developers.

Gemini Robotics and the Future of On-Device Embodied AI

The broader Gemini Robotics initiative has focused on unifying perception, reasoning, and action in physical environments. This on-device release bridges the gap between foundational AI research and deployable systems that can function autonomously in the real world.

While large VLA models like Gemini 1.5 have demonstrated impressive generalization across modalities, their inference latency and cloud dependency have limited their applicability in robotics. The on-device version addresses these limitations with optimized compute graphs, model compression, and task-specific architectures tailored for embedded GPUs.

Broader Implications for Robotics and AI Deployment

By decoupling powerful AI models from the cloud, Gemini Robotics On-Device paves the way for scalable, privacy-preserving robotics. It aligns with a growing trend toward edge AI, where computational workloads are shifted closer to data sources. This not only enhances safety and responsiveness but also ensures that robotic agents can operate in environments with strict latency or privacy requirements.

As DeepMind continues to broaden access to its robotics stack—including opening up its simulation platform and releasing benchmarks—researchers worldwide are now better equipped to experiment, iterate, and build reliable, real-time robotic systems.


Check out the Paper and Technical details. All credit for this research goes to the researchers of this project. Also, feel free to follow us on Twitter and don’t forget to join our 100k+ ML SubReddit and Subscribe to our Newsletter.


Asif Razzaq is the CEO of Marktechpost Media Inc.. As a visionary entrepreneur and engineer, Asif is committed to harnessing the potential of Artificial Intelligence for social good. His most recent endeavor is the launch of an Artificial Intelligence Media Platform, Marktechpost, which stands out for its in-depth coverage of machine learning and deep learning news that is both technically sound and easily understandable by a wide audience. The platform boasts of over 2 million monthly views, illustrating its popularity among audiences.



Source link

Share

Leave a comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Related Articles
Build a Low-Footprint AI Coding Assistant with Mistral Devstral
OpenAI

Build a Low-Footprint AI Coding Assistant with Mistral Devstral

In this Ultra-Light Mistral Devstral tutorial, a Colab-friendly guide is provided that...

ByteDance Researchers Introduce Seed-Coder: A Model-Centric Code LLM Trained on 6 Trillion Tokens
OpenAI

ByteDance Researchers Introduce Seed-Coder: A Model-Centric Code LLM Trained on 6 Trillion Tokens

Reframing Code LLM Training through Scalable, Automated Data Pipelines Code data plays...

ByteDance Researchers Introduce VGR: A Novel Reasoning Multimodal Large Language Model (MLLM) with Enhanced Fine-Grained Visual Perception Capabilities
OpenAI

ByteDance Researchers Introduce VGR: A Novel Reasoning Multimodal Large Language Model (MLLM) with Enhanced Fine-Grained Visual Perception Capabilities

Why Multimodal Reasoning Matters for Vision-Language Tasks Multimodal reasoning enables models to...

A Coding Implementation for Creating, Annotating, and Visualizing Complex Biological Knowledge Graphs Using PyBEL
OpenAI

A Coding Implementation for Creating, Annotating, and Visualizing Complex Biological Knowledge Graphs Using PyBEL

In this tutorial, we explore how to leverage the PyBEL ecosystem to...