Peeking inside the Black-Box: Reinforcement Learning for Explainable and Accurate Relation Extraction
Keywords: Relation Extraction, Large Language Models, Explainable AI
TL;DR: We propose Cognitive-Structured RE, a framework that improves relation extraction accuracy and explainability through a novel Hit@Dict RL reward.
Abstract: This paper introduces a framework for relation extraction (RE) that enhances both accuracy and explainability. The framework has two key components: (i) a reasoning mechanism that formulates relation extraction as a series of text-processing steps inspired by cognitive science, and (ii) an optimization process driven by reinforcement learning (RL) with a novel reward function designed to improve both task accuracy and explanation quality. We call our approach CogRE. Our framework addresses the lack of supervision for language-based explanations in traditional RE by promoting outputs that include important relation keywords. These keywords are drawn from a high-quality dictionary that is automatically constructed using an LLM. We evaluate our approach for the task of one-shot RE using two LLMs and two RE datasets. Our experiments show that CogRE improves explanation quality by addressing two common failure patterns in one-shot RE: poor attention focus and limited one-shot learning capability. For example, our cognitive-structured reasoning with Qwen2.5-15B-Instruct on One-shot NYT29 achieves 24.65\% F1, surpassing prior reasoning-based designs. Optimizing this approach with RL using our reward further improves performance by +23.46\% (absolute). Finally, human evaluation shows that our best model generates relational keywords closely aligned with gold labels, increasing human explanation quality ratings by 54\% (relative).
Supplementary Material: zip
Primary Area: interpretability and explainable AI
Submission Number: 24155
Loading