Introducing Symmetries to Black Box Meta Reinforcement LearningDownload PDF

Published: 10 Dec 2021, Last Modified: 05 May 2023NeurIPS 2021 Workshop MetaLearn PosterReaders: Everyone
Keywords: meta learning, reinforcement learning, black box, general purpose, generalisation, symmetries, invariance
TL;DR: We add symmetries such as permutation invariance to black-box meta reinforcement learners to increase their generalisation capabilities.
Abstract: Meta reinforcement learning (RL) attempts to discover new RL algorithms automatically from environment interaction. In so-called black-box approaches, the policy and the learning algorithm are jointly represented by a single neural network. These methods are very flexible, but they tend to underperform in terms of generalisation to new, unseen environments. In this paper, we explore the role of symmetries in meta-generalisation. We show that a recent successful meta RL approach that meta-learns an objective for backpropagation-based learning exhibits certain symmetries (specifically the reuse of the learning rule, and invariance to input and output permutations) that are not present in typical black-box meta RL systems. We hypothesise that these symmetries can play an important role in meta-generalisation. Building off recent work in black-box supervised meta learning, we develop a black-box meta RL system that exhibits these same symmetries. We show through careful experimentation that incorporating these symmetries can lead to algorithms with a greater ability to generalise to unseen action & observation spaces, tasks, and environments.
Contribution Process Agreement: Yes
Poster Session Selection: Poster session #1 (12:00 UTC), Poster session #2 (15:00 UTC), Poster session #3 (16:50 UTC)
0 Replies