Jump to content

Gemini Robotics

From Wikipedia, the free encyclopedia

Gemini Robotics is an advanced vision-language-action model developed by Google DeepMind[1] in partnership with Apptronik.[2] It is based on the Gemini 2.0 large language model.[3] It is tailored for robotics applications and can understand new situations.[4][5] There is a related version called Gemini Robotics-ER, which stands for embodied reasoning.[3] The two models were launched on March 12, 2025.[5]

On June 24, 2025, Google DeepMind released Gemini Robotics On-Device, a variant designed and optimized to run locally on robotic devices.[6]

Access to Gemini Robotics models is currently restricted to trusted testers, including Agile Robots, Agility Robots, Boston Dynamics, and Enchanted Tools.[2]

References

[edit]
  1. ^ "Gemini Robotics". deepmind.google. Retrieved March 12, 2025.
  2. ^ a b Parada, Carolina. "Gemini Robotics brings AI into the physical world". Google DeepMind. Retrieved July 11, 2025.
  3. ^ a b Knight, Will (May 12, 2025). "Google's Gemini Robotics AI Model Reaches Into the Physical World". WIRED. Retrieved March 12, 2025.
  4. ^ "Google introduces new AI models for rapidly growing robotics industry". Reuters. March 12, 2025. Retrieved March 12, 2025.
  5. ^ a b Roth, Emma (March 12, 2025). "Google DeepMind's new AI models help robots perform physical tasks, even without training". The Verge. Retrieved March 12, 2025.
  6. ^ Parada, Carolina. "Gemini Robotics On-Device brings AI to local robotic devices". Google DeepMind. Retrieved July 11, 2025.