Class-Conditional Conformal Prediction for Imbalanced Data via Top-$k$ Classes

23 Sept 2023 (modified: 11 Feb 2024)Submitted to ICLR 2024EveryoneRevisionsBibTeX
Primary Area: general machine learning (i.e., none of the above)
Code Of Ethics: I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics.
Keywords: Uncertainty Quantification, Conformal Prediction, Imbalanced Data, Class-conditional Coverage, Deep Models
Submission Guidelines: I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2024/AuthorGuide.
TL;DR: Provable conformal prediction method to produce small prediction sets for a target class-conditional coverage for imbalanced classification problems using the top-k labels from classifier scores.
Abstract: Classification tasks where data contains skewed class proportions (aka {\em imbalanced data}) arises in many real-world applications including medical diagnosis. Safe deployment of classifiers for imbalanced data settings require theoretically-sound uncertainty quantification. Conformal prediction (CP) is a promising framework for producing prediction sets from black-box classifiers with a user-specified coverage (i.e., true class is contained with high probability). Existing class-conditional CP (CCP) method employs a black-box classifier to find one threshold for each class during calibration and then includes every class label that meets the corresponding threshold for testing inputs, leading to large prediction sets. This paper studies the problem of how to develop provable CP methods with small prediction sets for the class-conditional coverage setting and makes several contributions. First, we theoretically show that marginal CP can perform arbitrarily poorly and cannot provide coverage guarantee for minority classes. Second, we propose a principled algorithm referred to as {\em $k$-Class-conditional CP ($k$-CCP)}. The key idea behind $k$-CCP is to restrict the candidate labels for the prediction set of a testing input to only top-$k$ labels based on the classifier scores (in contrast to all labels in CCP). Third, we prove that $k$-CCP provides class-conditional coverage and produces smaller prediction sets over the CCP method. Our experiments on benchmark datasets demonstrate that $k$-CCP achieves class-conditional coverage and produces smaller prediction sets over baseline methods.
Anonymous Url: I certify that there is no URL (e.g., github page) that could be used to find authors' identity.
Supplementary Material: zip
No Acknowledgement Section: I certify that there is no acknowledgement section in this submission for double blind review.
Submission Number: 8076
Loading