Learning to Infer Unseen Contexts in Causal Contextual Reinforcement LearningDownload PDF

Anonymous

Published: 15 Jun 2022, Last Modified: 05 May 2023SSL-RL 2021 PosterReaders: Everyone
Keywords: Contextual Reinforcement Learning, Causal Reinforcement Learning, Disentangled Representations, Contextual Attention
TL;DR: We introduce a new task to evaluate reasoning capabilities of contextual agents, showing non-causal agents struggle to generalise in this task. We introduce a contextual attention module that helps agents discover causal factors & better generalise.
Abstract: In Contextual Reinforcement Learning (CRL), a change in the context variable can cause a change in the distribution of the states. Hence contextual agents must be able to learn adaptive policies that can change when a context changes. Furthermore, in certain scenarios agents have to deal with unseen contexts, and be able to choose suitable actions. In order to generalise onto unseen contexts, agents need to not only detect and adapt to previously observed contexts, but also reason about how a context is constructed, and what are the causal factors of context variables. In this paper, we propose a new task and environment for Causal Contextual Reinforcement Learning (CCRL), where the performance of different agents can be compared in a causal reasoning task. Furthermore, we introduce a Contextual Attention Module that allows the agent to incorporate disentangled features as the contextual factors, which results in performance improvement of the agent in unseen contexts. Finally, we demonstrate that non-causal agents fail to generalise onto unseen contexts, while the agents incorporating the proposed module can achieve better performance in unseen contexts.
0 Replies

Loading