Gemini Robotics

Last updated

Gemini Robotics is an advanced vision-language-action model developed by Google DeepMind [1] in partnership with Apptronik. [2] It is based on the Gemini 2.0 large language model. [3] It is tailored for robotics applications and can understand new situations. [4] [5] There is a related version called Gemini Robotics-ER, which stands for embodied reasoning. [3] The two models were launched on March 12, 2025. [5]

On June 24, 2025, Google DeepMind released Gemini Robotics On-Device, a variant designed and optimized to run locally on robotic devices. [6]

Access to Gemini Robotics models is currently restricted to trusted testers, including Agile Robots, Agility Robots, Boston Dynamics, and Enchanted Tools. [2]

References

  1. "Gemini Robotics". deepmind.google. Retrieved March 12, 2025.
  2. 1 2 Parada, Carolina. "Gemini Robotics brings AI into the physical world". Google DeepMind. Retrieved July 11, 2025.
  3. 1 2 Knight, Will (May 12, 2025). "Google's Gemini Robotics AI Model Reaches Into the Physical World". WIRED . Retrieved March 12, 2025.
  4. "Google introduces new AI models for rapidly growing robotics industry". Reuters . March 12, 2025. Retrieved March 12, 2025.
  5. 1 2 Roth, Emma (March 12, 2025). "Google DeepMind's new AI models help robots perform physical tasks, even without training". The Verge . Retrieved March 12, 2025.
  6. Parada, Carolina. "Gemini Robotics On-Device brings AI to local robotic devices". Google DeepMind. Retrieved July 11, 2025.