Intervention Design for Causal Representation LearningDownload PDF

10 Jun 2022, 18:37 (modified: 09 Jul 2022, 18:32)CRL@UAI 2022 PosterReaders: Everyone
Keywords: Intervention Design, Experiment Design, Causal Representation Learning
TL;DR: We derive a minimal bound of experiments that guarantee identifiability of causal variables, opening up new opportunities for using intervention design for causal representation learning.
Abstract: In this paper, we take a first step towards bringing two fields of causality closer together: intervention design and causal representation learning. Intervention design is a well studied task in classic causal discovery, which aims at finding the minimal sets of experiments under which the causal graph can be identified. Causal representation learning aims at recovering causal variables from high-dimensional entangled observations. In recent work in causal representation, interventions are exploited to improve identifiability, similarly to classic causal discovery. Hence, the same task becomes relevant in this setting as well: how many experiments are minimally needed to identify the latent causal variables? Based on the recent causal representation learning method CITRIS, we show that for $K$ causal variables, $\lfloor \log_2 (K) \rfloor + 2$ experiments are sufficient to identify causal variables from temporal, intervened sequences, which is only one more experiment than needed for classic causal discovery in the worst case. Further, we show that this bound holds empirically in experiments on a 3D rendered video dataset.
4 Replies

Loading