L-C2ST: Local Diagnostics for Posterior Approximations in Simulation-Based Inference

Published: 21 Sept 2023, Last Modified: 23 Jan 2024NeurIPS 2023 posterEveryoneRevisionsBibTeX
Keywords: machine learning, calibration, simulation-based inference, neuroscience, normalizing flows, classifier two-sample tests
TL;DR: A validation method for conditional density estimators particularly tailored for the simulation-based inference setting, with theoretical guarantees and interpretable diagnostics.
Abstract: Many recent works in simulation-based inference (SBI) rely on deep generative models to approximate complex, high-dimensional posterior distributions. However, evaluating whether or not these approximations can be trusted remains a challenge. Most approaches evaluate the posterior estimator only in expectation over the observation space. This limits their interpretability and is not sufficient to identify for which observations the approximation can be trusted or should be improved. Building upon the well-known classifier two-sample test (C2ST), we introduce $\ell$-C2ST, a new method that allows for a local evaluation of the posterior estimator at any given observation. It offers theoretically grounded and easy to interpret -- e.g. graphical -- diagnostics, and unlike C2ST, does not require access to samples from the true posterior. In the case of normalizing flow-based posterior estimators, $\ell$-C2ST can be specialized to offer better statistical power, while being computationally more efficient. On standard SBI benchmarks, $\ell$-C2ST provides comparable results to C2ST and outperforms alternative local approaches such as coverage tests based on highest predictive density (HPD). We further highlight the importance of local evaluation and the benefit of interpretability of $\ell$-C2ST on a challenging application from computational neuroscience.
Supplementary Material: zip
Submission Number: 6750
Loading