Look where you look! Saliency-guided Q-networks for generalization in visual Reinforcement LearningDownload PDF

Published: 31 Oct 2022, 18:00, Last Modified: 24 Oct 2022, 21:25NeurIPS 2022 AcceptReaders: Everyone
Keywords: Reinforcement learning, Generalization
TL;DR: We present a generic method improving generalization for visual reinforcement learning based on attribution maps.
Abstract: Deep reinforcement learning policies, despite their outstanding efficiency in simulated visual control tasks, have shown disappointing ability to generalize across disturbances in the input training images. Changes in image statistics or distracting background elements are pitfalls that prevent generalization and real-world applicability of such control policies. We elaborate on the intuition that a good visual policy should be able to identify which pixels are important for its decision, and preserve this identification of important sources of information across images. This implies that training of a policy with small generalization gap should focus on such important pixels and ignore the others. This leads to the introduction of saliency-guided Q-networks (SGQN), a generic method for visual reinforcement learning, that is compatible with any value function learning method. SGQN vastly improves the generalization capability of Soft Actor-Critic agents and outperforms existing state-of-the-art methods on the Deepmind Control Generalization benchmark, setting a new reference in terms of training efficiency, generalization gap, and policy interpretability.
Supplementary Material: zip
17 Replies