\textit{Dial BeInfo for Faithfulness}: Improving Factuality of Information-Seeking Dialogue via Behavioural Fine-Tuning

ACL ARR 2024 April Submission115 Authors

13 Apr 2024 (modified: 27 May 2024)ACL ARR 2024 April SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Abstract: Factual faithfulness is a crucial requirement in information-seeking dialogue: the system should respond to the user queries so that the responses are meaningful and aligned with the knowledge provided to the system. However, most modern large language models (LLMs) suffer from hallucinations, that is, they generate responses not supported by or even contradicting the knowledge source. To mitigate the issue and increase faithfulness of information-seeking dialogue systems supported by the LLMs, we introduce BEINFO, a simple yet effective method that applies ‘behavioural tuning’ on the LLMs to aid information-seeking dialogue. Relying on three standard information seeking dialogue datasets, we show that models tuned with BEINFO become considerably more faithful to the knowledge source both for datasets and domains seen during BEINFO-tuning, as well as on unseen domains, when applied in a zero-shot manner. In addition, we present a ‘real-life’ case study on conversations with real users, showcasing that the models with 3B parameters (e.g., Flan-T5) tuned with BEINFO demonstrate strong performance on data from real ‘production’ conversations: when tuned on a limited amount of such realistic in-domain dialogues, they surpass much larger LLMs used ‘off-the-shelf’, both on automatic and human evaluation metrics.
Paper Type: Long
Research Area: Dialogue and Interactive Systems
Research Area Keywords: Information-Seeking Dialogue;Factuality
Contribution Types: NLP engineering experiment
Languages Studied: English
Submission Number: 115
Loading

OpenReview is a long-term project to advance science through improved peer review with legal nonprofit status. We gratefully acknowledge the support of the OpenReview Sponsors. © 2025 OpenReview