Keywords: clinical notes, distribution shift, shift explanation, benchmark, interpretability, large language models
Track: Proceedings
Abstract: Distribution shift can lead to degradation in the performance of machine learning models. This concern is particularly salient in medicine, in which several forces can lead to shifts in Electronic Health Record (EHR) data. Distribution shift in the text domain is vastly understudied, but increasingly important, given the widespread integration of large language models into clinical workflows. Identifying the existence of a shift is necessary but insufficient; actionability often requires understanding the nature of the shift. To address this challenge, we establish an extensible benchmark suite that induces synthetic distribution shifts using real clinical notes and develop two methods to assess generated shift explanations. We further introduce SIReNs, a general-domain end-to-end approach that explains distributional differences between two datasets by selecting representative notes from each. The SIReNs method was evaluated on both binary and continuous feature shifts, and the results show that it recovers salient binary shifts well, but struggles with more subtle shifts. A substantial gap remains to a ground-truth oracle for continuous shifts, suggesting room for improvement in future methods.
General Area: Applications and Practice
Specific Subject Areas: Explainability & Interpretability, Evaluation Methods & Validity, Natural Language Processing, Dataset Release & Characterization
Data And Code Availability: Yes
Ethics Board Approval: No
Entered Conflicts: I confirm the above
Anonymity: I confirm the above
Submission Number: 133
Loading