Pointwise Information Measures as Confidence Estimators in Deep Neural Networks: A Comparative Study

Published: 01 May 2025, Last Modified: 18 Jun 2025ICML 2025 posterEveryoneRevisionsBibTeXCC BY 4.0
Abstract: Estimating the confidence of deep neural network predictions is crucial for safe deployment in high-stakes applications. While softmax probabilities are commonly used, they are often poorly calibrated, and existing calibration methods have been shown to be detrimental to failure prediction. In this paper, we propose using information-theoretic measures to estimate prediction confidence in a post-hoc manner, without modifying network architecture or training. Specifically, we compare three pointwise information (PI) measures: pointwise mutual information (PMI), pointwise V-information (PVI), and the recently proposed pointwise sliced mutual information (PSI). These measures are theoretically grounded in their relevance to predictive uncertainty, with properties such as invariance, convergence rates, and sensitivity to geometric attributes like margin and intrinsic dimensionality. Through extensive experiments on benchmark computer vision models and datasets, we find that PVI consistently outperforms or matches existing baselines in post-hoc confidence estimation, excelling over PMI and PSI in both failure prediction and calibration. This aligns with our theoretical insights, which suggest that PVI offers the most balanced trade-offs. Finally, we show that replacing softmax outputs with PVI in existing confidence estimation methods can further enhance performance.
Lay Summary: Deep neural networks often produce overconfident predictions, which can be dangerous in high-stakes applications like healthcare or autonomous driving. While softmax probabilities are commonly used to estimate confidence, they are often poorly calibrated, and many existing methods actually make failure prediction worse. In this work, we explore three information-theoretic measures: (1) Pointwise Mutual Information (PMI), (2) Pointwise V-Information (PVI), and (3) Pointwise Sliced Mutual Information (PSI), as post-hoc tools for estimating confidence, without changing the model’s architecture or training. We analyze how these measures behave under different data transformations and geometric properties like margin and dimensionality. Our experiments on standard computer vision benchmarks show that PVI consistently outperforms PMI, PSI, and existing confidence estimation methods. These findings suggest that PVI offers a more balanced and reliable way to assess model confidence, especially for detecting when the model is likely to fail.
Link To Code: https://github.com/kentridgeai/PI-Conf-Est
Primary Area: Social Aspects->Accountability, Transparency, and Interpretability
Keywords: information theory, confidence estimation, deep neural networks
Submission Number: 14884
Loading