Abstract: Large Language Models require both contextual knowledge and parametric memory, but these sources can disagree.
Prior investigations on contextual question answering tasks report a preference toward parametric knowledge under conflict, yet they focus almost exclusively on tasks that should always rely on the given passage, leaving open how this behavior manifests when tasks demand different amounts and kinds of knowledge.
We study this question with a model-agnostic diagnostic framework that (i) automatically detects disagreements between a model’s beliefs and a curated knowledge set, and (ii) injects controlled conflicts into tasks.
The resulting datasets span two orthogonal dimensions: task knowledge reliance and conflict plausibility.
Evaluating representative open-source LLMs, we find that:
(1) performance degradation from conflict correlates with a task’s knowledge reliance;
(2) explanatory rationales and simple reiteration both increase context reliance—helpful for context-only tasks but harmful when parametric knowledge should dominate;
(3) These behaviors raise concerns about the validity of model-based evaluation and underscore the need to account for knowledge conflict in the deployment of LLMs.
Paper Type: Long
Research Area: Interpretability and Analysis of Models for NLP
Research Area Keywords: Knowledge Conflict, Context-memory Conflict
Contribution Types: Model analysis & interpretability, Data resources
Languages Studied: English
Submission Number: 819
Loading