On the properties of variational approximations of Gibbs posteriors
Abstract: The PAC-Bayesian approach is a powerful set of techniques to derive non-asymptoticrisk bounds for random estimators. The corresponding optimal distribution of estimators,usually called the Gibbs posterior, is unfortunately often intractable. One may samplefrom it using Markov chain Monte Carlo, but this is usually too slow for big datasets.We consider instead variational approximations of the Gibbs posterior, which are fastto compute. We undertake a general study of the properties of such approximations.Our main finding is that such a variational approximation has often the same rate ofconvergence as the original PAC-Bayesian procedure it approximates. In addition, weshow that, when the risk function is convex, a variational approximation can be obtainedin polynomial time using a convex solver. We give finite sample oracle inequalities for thecorresponding estimator. We specialize our results to several learning tasks (classification,ranking, matrix completion), discuss how to implement a variational approximation ineach case, and illustrate the good properties of said approximation on real datasets.
0 Replies
Loading