Introducing Symmetries to Black Box Meta Reinforcement LearningDownload PDF

30 Sept 2021, 15:22 (edited 01 Dec 2021)NeurIPS 2021 Workshop MetaLearn PosterReaders: Everyone
  • Keywords: meta learning, reinforcement learning, black box, general purpose, generalisation, symmetries, invariance
  • TL;DR: We add symmetries such as permutation invariance to black-box meta reinforcement learners to increase their generalisation capabilities.
  • Abstract: Meta reinforcement learning (RL) attempts to discover new RL algorithms automatically from environment interaction. In so-called black-box approaches, the policy and the learning algorithm are jointly represented by a single neural network. These methods are very flexible, but they tend to underperform in terms of generalisation to new, unseen environments. In this paper, we explore the role of symmetries in meta-generalisation. We show that a recent successful meta RL approach that meta-learns an objective for backpropagation-based learning exhibits certain symmetries (specifically the reuse of the learning rule, and invariance to input and output permutations) that are not present in typical black-box meta RL systems. We hypothesise that these symmetries can play an important role in meta-generalisation. Building off recent work in black-box supervised meta learning, we develop a black-box meta RL system that exhibits these same symmetries. We show through careful experimentation that incorporating these symmetries can lead to algorithms with a greater ability to generalise to unseen action & observation spaces, tasks, and environments.
  • Contribution Process Agreement: Yes
  • Poster Session Selection: Poster session #1 (12:00 UTC), Poster session #2 (15:00 UTC), Poster session #3 (16:50 UTC)
0 Replies