Catch & carry: reusable neural controllers for vision-guided whole-body tasks

J Merel, S Tunyasuvunakool, A Ahuja, Y Tassa… - ACM Transactions on …, 2020 - dl.acm.org
ACM Transactions on Graphics (TOG), 2020dl.acm.org
We address the longstanding challenge of producing flexible, realistic humanoid character
controllers that can perform diverse whole-body tasks involving object interactions. This
challenge is central to a variety of fields, from graphics and animation to robotics and motor
neuroscience. Our physics-based environment uses realistic actuation and first-person
perception-including touch sensors and egocentric vision-with a view to producing active-
sensing behaviors (eg gaze direction), transferability to real robots, and comparisons to the …
We address the longstanding challenge of producing flexible, realistic humanoid character controllers that can perform diverse whole-body tasks involving object interactions. This challenge is central to a variety of fields, from graphics and animation to robotics and motor neuroscience. Our physics-based environment uses realistic actuation and first-person perception - including touch sensors and egocentric vision - with a view to producing active-sensing behaviors (e.g. gaze direction), transferability to real robots, and comparisons to the biology. We develop an integrated neural-network based approach consisting of a motor primitive module, human demonstrations, and an instructed reinforcement learning regime with curricula and task variations. We demonstrate the utility of our approach for several tasks, including goal-conditioned box carrying and ball catching, and we characterize its behavioral robustness. The resulting controllers can be deployed in real-time on a standard PC.1
ACM Digital Library