A Pilot Study on Doubt Robustness of LLMs in Clinical Prediction Explanation

Published: 02 Mar 2026, Last Modified: 03 Mar 2026ICLR 2026 Workshop ICBINBEveryoneRevisionsCC BY 4.0
Keywords: Large Language Model, Explainable AI, Robustness, Clinical Prediction
TL;DR: We demonstrate that simple doubt prompts can easily destabilize LLM-generated clinical explanations.
Abstract: We study large language models (LLMs) as clinical explanation generators and evaluate their robustness to user doubt in interactive settings. Using an in-hospital mortality prediction task on the MIMIC-III dataset, we examine how simple challenge prompts affect the consistency of LLM-generated explanations. We adopt the concept of doubt robustness and assess it by prompting models to explain risk predictions and indicate agreement, followed by doubt-inducing queries. Our results show that instruction-tuned models frequently reverse their initial stance, while reasoning-enhanced models exhibit improved but still limited stability. Further analysis suggests that LLMs rely heavily on model outputs rather than ground-truth labels, reducing explanation faithfulness. These findings highlight the need for robustness-oriented evaluation of clinical explanation systems.
Email Sharing: We authorize the sharing of all author emails with Program Chairs.
Data Release: We authorize the release of our submission and author names to the public in the event of acceptance.
Submission Number: 47
Loading