Functional Wasserstein Bridge Inference for Bayesian Deep Learning

Published: 26 Apr 2024, Last Modified: 15 Jul 2024UAI 2024 oralEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Bayesian deep learning, functional variational inference, Wasserstein distance
Abstract: Bayesian deep learning (BDL) is an emerging field that combines the strong function approximation power of deep learning with the uncertainty modeling capabilities of Bayesian methods. In addition to those virtues, however, there are accompanying issues brought by such a combination to the classical parameter-space variational inference, such as the nonmeaningful priors, intricate posteriors, and possible pathologies. In this paper, we propose a new function-space variational inference solution called Functional Wasserstein Bridge Inference (FWBI), which can assign meaningful functional priors and obtain well-behaved posterior. Specifically, we develop a Wasserstein distance-based bridge to avoid the potential pathological behaviors of Kullback–Leibler (KL) divergence between stochastic processes that arise in most existing functional variational inference approaches. The derived functional variational objective is well-defined and proved to be a lower bound of the model evidence. We demonstrate the improved predictive performance and better uncertainty quantification of our FWBI on several tasks compared with various parameter-space and function-space variational methods.
Supplementary Material: zip
List Of Authors: Wu, Mengjing and Xuan, Junyu and Lu, Jie
Latex Source Code: zip
Signed License Agreement: pdf
Submission Number: 214
Loading