Evaluating Transparency of Machine Generated Fact Checking Explanations

ACL ARR 2024 June Submission1648 Authors

14 Jun 2024 (modified: 10 Aug 2024)ACL ARR 2024 June SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Abstract: An important factor when it comes to generating fact-checking explanations is the selection of evidence: intuitively, high-quality explanations can only be generated given the right evidence. In this work, we investigate the impact of human-curated vs. machine-selected evidence for explanation generation using large language models. To assess the quality of explanations, we focus on transparency (whether an explanation cites sources properly) and utility (whether an explanation is helpful in clarifying a claim). Surprisingly, we found that large language models generate similar or higher quality explanations using machine-selected evidence, suggesting carefully curated evidence (by humans) may not be necessary. That said, even with the best model, the generated explanations are not always faithful to the sources, suggesting further room for improvement in explanation generation for fact-checking.
Paper Type: Short
Research Area: NLP Applications
Research Area Keywords: fact checking, text-to-text generation, free-text/natural language explanations, prompting, evaluation
Contribution Types: NLP engineering experiment, Data analysis
Languages Studied: English
Submission Number: 1648
Loading