SConU: Selective Conformal Uncertainty in Large Language Models

ACL ARR 2025 February Submission702 Authors

10 Feb 2025 (modified: 09 May 2025)ACL ARR 2025 February SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Abstract: As large language models are increasingly utilized in real-world applications, guarantees of task-specific performance are essential for their reliable deployment. Recent studies have introduced various conformal uncertainty criteria grounded in split conformal prediction, which offer user-specified correctness coverage. However, existing frameworks often fail to identify uncertainty data outliers that violate the exchangeability assumption, leading to unbounded miscoverage rates and unactionable prediction sets. In this paper, we propose a novel approach termed Selective Conformal Uncertainty (SConU), which, for the first time, implements significance tests, by developing two conformal p-values that are instrumental in determining whether a given sample deviates from the uncertainty distribution of the calibration set at a specific manageable risk level. Our approach not only facilitates rigorous management of miscoverage rates across both single-domain and interdisciplinary contexts, but also enhances the efficiency of predictions. Furthermore, we comprehensively analyze the components of the conformal procedures, aiming to approximate conditional coverage, particularly in high-stakes question-answering tasks.
Paper Type: Long
Research Area: Question Answering
Research Area Keywords: inference methods,statistical testing for evaluation
Contribution Types: Theory
Languages Studied: English
Submission Number: 702
Loading