Track: full paper
Keywords: Imitation Learning, Robot Manipulation, Computer Vision
TL;DR: Our new highly efficient imitation learning paradigm enables teaching a robot 1000 distinct tasks in just 24 hours.
Abstract: Humans are remarkably efficient at learning tasks from demonstrations, but today's imitation learning methods for robot manipulation often require hundreds or thousands of demonstrations per task. To bridge this gap, we discovered that decomposing reasoning into two sequential phases – object alignment and then object interaction – can enable robots to learn everyday tasks from just a single demonstration. We systematically evaluated this decomposition by comparing different design choices for each phase of reasoning, and by studying the generalisation and scaling trends with respect to today’s dominant paradigm of behavioural cloning with a single-phase monolithic policy. Through 3,450 real-world policy rollouts, we found compelling conclusions that, focussing on efficient learning from few demonstrations per task, decomposition significantly outperforms learning the full trajectory in a single phase, and for each phase, reasoning via retrieval in a learned latent space outperforms behavioural cloning. Building on these insights, we then designed Multi-Task Trajectory Transfer (MT3), a novel imitation learning method based on decomposition and retrieval which is capable of learning everyday manipulation tasks from only a single demonstration each, whilst also generalising efficiently to novel objects. We found that this major leap in data efficiency ultimately enabled us to teach a robot 1000 distinct everyday tasks within just 24 hours of human demonstrator time. Videos of our experiments can be found on our website https://www.robot-learning.uk/learning-1000-tasks.
Anonymization: This submission has been anonymized for double-blind review via the removal of identifying information such as names, affiliations, and identifying URLs.
Presenter: ~Kamil_Dreczkowski1
Format: Yes, the presenting author will attend in person if this work is accepted to the workshop.
Funding: Yes, the presenting author of this submission falls under ICLR’s funding aims, and funding availability would significantly influence their ability to attend the workshop in person.
Submission Number: 70
Loading