Representation Learning for Out-of-distribution Generalization in Reinforcement LearningDownload PDF

Published: 22 Jul 2021, Last Modified: 08 Sept 2024URL 2021 PosterReaders: Everyone
Keywords: representations, robustness, out-of-distribution, generalization, reinforcement learning
TL;DR: We present a large-scale empirical analysis studying the out-of-distribution generalization of RL control policies from pre-trained representations.
Abstract: Learning data representations that are useful for various downstream tasks is a cornerstone of artificial intelligence. While existing methods are typically evaluated on downstream tasks such as classification or generative image quality, we propose to assess representations through their usefulness in downstream control tasks, such as reaching or pushing objects. By training over 10,000 reinforcement learning policies, we extensively evaluate to what extent different representation properties affect out-of-distribution (OOD) generalization. Finally, we demonstrate zero-shot transfer of these policies from simulation to the real world, without any domain randomization or fine-tuning. This paper aims to establish the first systematic characterization of the usefulness of learned representations for real-world OOD downstream tasks.
Community Implementations: [![CatalyzeX](/images/catalyzex_icon.svg) 1 code implementation](https://www.catalyzex.com/paper/representation-learning-for-out-of/code)
1 Reply

Loading