How to deal with missing data in supervised deep learning?Download PDF

Jun 10, 2020 (edited Jul 12, 2020)ICML 2020 Workshop Artemiss SubmissionReaders: Everyone
  • Keywords: missing data, supervised learning, deep learning
  • TL;DR: joint DLVM and discriminative model leads to performance gains in deep supervised learning with missing values
  • Abstract: The issue of missing data in supervised learning has been largely overlooked, especially in the deep learning community. We investigate strategies to adapt neural architectures to handle missing values. Here, we focus on regression and classification problems where the features are assumed to be missing at random. Of particular interest are schemes that allow to reuse as-is a neural discriminative architecture. One scheme involves imputing the missing values with learnable constants. We propose a second novel approach that leverages recent advances in deep generative modelling. More precisely, a deep latent variable model can be learned jointly with the discriminative model, using importance-weighted variational inference in an end-to-end way. This hybrid approach, which mimics multiple imputation, also allows to impute the data, by relying on both the discriminative and the generative model. We also discuss ways of using a pre-trained generative model to train the discriminative one. In domains where powerful deep generative models are available, the hybrid approach leads to large performance gains.
2 Replies