Exploring the Generalizability of Factual Hallucination Mitigation via Enhancing Precise Knowledge Utilization

ACL ARR 2025 May Submission7370 Authors

20 May 2025 (modified: 03 Jul 2025)ACL ARR 2025 May SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Abstract: Large Language Models (LLMs) often struggle to align their responses with objective facts, resulting in the issue of $\textbf{factual hallucinations}$, which can be difficult to detect and mislead users without relevant knowledge. Although post-training techniques have been employed to mitigate the issue, existing methods usually suffer from poor generalization and trade-offs in different capabilities. In this paper, we propose to address it by directly augmenting LLM's fundamental ability to precisely leverage its knowledge and introduce $\textbf{PKUE}$, which fine-tunes the model on self-generated responses to precise and simple factual questions through preference optimization. Furthermore, we construct $\textbf{FactualBench}$, a comprehensive and precise factual QA dataset containing 181k Chinese data spanning 21 domains, to facilitate both evaluation and training. Extensive experiments demonstrate that PKUE significantly improves LLM overall performance, with consistent enhancement across factual tasks of various forms, general tasks beyond factuality, and tasks in a different language.
Paper Type: Long
Research Area: Question Answering
Research Area Keywords: generalization,factuality,question generation,NLP datasets
Contribution Types: NLP engineering experiment, Data resources
Languages Studied: Chinese,English
Submission Number: 7370
Loading