TINY: Semantic-based Uncertainty Quantification in LLMS: A Case Study on Medical Explanation Generation Task.

Published: 05 Mar 2025, Last Modified: 31 Mar 2025QUESTION PosterEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Uncertainty Quantification, Large Language Models, trustworthy AI, Interpretability, Natural Language Generation, Explanation Generation, Generative Models
TL;DR: The recently proposed Semantic Density framework is used to quantify uncertainty in medical explanation generation datasets, demonstrating the generalizability of the method to specialized domains without additional modifications.
Abstract: Given the often sensible and sometimes nonsensical outputs that modern Large Language Models (LLMs) generate, how should we interpret confident claims such as "Strawberry has two 'r's"? One tool that can be applied to such overconfident and hallucinatory claims is uncertainty quantification. In particular, this paper investigates a semantic density method to quantify uncertainty in LLM-generated medical explanations. Semantic density makes use of semantic similarity comparisons instead of lexical matching, and delivers per-response estimates of uncertainty. The results demonstrate that the semantic density framework remains performant when applied in specialized domains, and raises additional considerations around the utility of the ROUGE metric for semantic evaluations.
Submission Number: 29
Loading