Using Deep Learning to Bootstrap Abstractions for Robot PlanningDownload PDF

04 Mar 2022, 07:18 (edited 16 Apr 2022)ICLR 2022 GPL PosterReaders: Everyone
  • Keywords: Learning Abstractions for Planning, Deep Learning, Hierarchical Planning, Motion Planning, Learning for Motion Planning
  • TL;DR: This paper presents a framework that uses deep learning to learn hierarchical state and action abstractions and introduces a novel multi-source multi-directional hierarchical planning algorithm to efficiently use learned abstractions.
  • Abstract: This paper addresses the problem of learning abstractions that boost robot planning performance while providing strong guarantees of reliability. Although state-of-the-art hierarchical robot planning algorithms allow robots to efficiently compute long-horizon motion plans for achieving user desired tasks, these methods typically rely upon environment-dependent state and action abstractions that need to be hand-designed by experts. We present a new approach for bootstrapping the entire hierarchical planning process. This allows us to compute abstract states and actions for new environments automatically using the critical regions predicted by a deep neural network with an auto-generated robot-specific architecture. We show that the learned abstractions can be used with a novel multi-source bi-directional hierarchical robot planning algorithm that is sound and probabilistically complete. An extensive empirical evaluation on twenty different settings using holonomic and non-holonomic robots shows that (a) our learned abstractions provide the information necessary for efficient multi-source hierarchical planning; and that (b) this approach of learning, abstractions, and planning outperforms state-of-the-art baselines by nearly a factor of ten in terms of planning time on test environments not seen during training.
1 Reply