Joint Inference for Neural Network Depth and Dropout RegularizationDownload PDF

21 May 2021, 20:42 (edited 29 Oct 2021)NeurIPS 2021 PosterReaders: Everyone
  • Keywords: Deep Learning, Nonparametric Bayesian Model Selection, Stochastic Variational Inference
  • TL;DR: Joint inference for deep neural network depth and dropout regularization via Bayesian model selection
  • Abstract: Dropout regularization methods prune a neural network's pre-determined backbone structure to avoid overfitting. However, a deep model still tends to be poorly calibrated with high confidence on incorrect predictions. We propose a unified Bayesian model selection method to jointly infer the most plausible network depth warranted by data, and perform dropout regularization simultaneously. In particular, to infer network depth we define a beta process over the number of hidden layers which allows it to go to infinity. Layer-wise activation probabilities induced by the beta process modulate neuron activation via binary vectors of a conjugate Bernoulli process. Experiments across domains show that by adapting network depth and dropout regularization to data, our method achieves superior performance comparing to state-of-the-art methods with well-calibrated uncertainty estimates. In continual learning, our method enables neural networks to dynamically evolve their depths to accommodate incrementally available data beyond their initial structures, and alleviate catastrophic forgetting.
  • Supplementary Material: pdf
  • Code Of Conduct: I certify that all co-authors of this work have read and commit to adhering to the NeurIPS Statement on Ethics, Fairness, Inclusivity, and Code of Conduct.
  • Code: https://github.com/kckishan/Depth_and_Dropout
15 Replies

Loading