Navigation
Recherche
|
Google's New Robot AI Can Fold Delicate Origami, Close Zipper Bags
mercredi 12 mars 2025, 21:50 , par Slashdot
![]() In 2023, we covered Google's RT-2, which represented a notable step toward more generalized robotic capabilities by using Internet data to help robots understand language commands and adapt to new scenarios, then doubling performance on unseen tasks compared to its predecessor. Two years later, Gemini Robotics appears to have made another substantial leap forward, not just in understanding what to do but in executing complex physical manipulations that RT-2 explicitly couldn't handle. While RT-2 was limited to repurposing physical movements it had already practiced, Gemini Robotics reportedly demonstrates significantly enhanced dexterity that enables previously impossible tasks like origami folding and packing snacks into Zip-loc bags. This shift from robots that just understand commands to robots that can perform delicate physical tasks suggests DeepMind may have started solving one of robotics' biggest challenges: getting robots to turn their 'knowledge' into careful, precise movements in the real world. DeepMind claims Gemini Robotics 'more than doubles performance on a comprehensive generalization benchmark compared to other state-of-the-art vision-language-action models.' Google is advancing this effort through a partnership with Apptronik to develop next-generation humanoid robots powered by Gemini 2.0. Availability timelines or specific commercial applications for the new AI models were not made available. Read more of this story at Slashdot.
https://hardware.slashdot.org/story/25/03/12/2018222/googles-new-robot-ai-can-fold-delicate-origami-...
Voir aussi |
56 sources (32 en français)
Date Actuelle
ven. 14 mars - 18:44 CET
|