Abstract: Large Language Models (LLMs) are effective at natural language reasoning but still struggle with answering commonsense questions that require implicit knowledge of the world. LLMs rely on knowledge learned through training, which can be limited to specific domains and may use inaccurate knowledge, resulting in hallucinations. To alleviate these, recent research integrates external knowledge sources (e.g., Fine-tuning, Self-revision, Retrieval-Augmented Generation (RAG), and Chain-of-Thought (CoT)). However, regular CoT reasoning merely presents the answering process in a specious form, where individual steps are challenging to verify. In this paper, we propose a novel approach called Revisable Chain-of-Thought to address an important commonsense question answering task, the Winograd Schema Challenge. Inspired by the cognitive logic of ``rising from the abstract to the concrete,'' Revisable CoT decomposes knowledge into three distinct categories: meta-knowledge, transfer knowledge, and instantiated knowledge, each handled in separate steps. This framework emphasizes step-by-step verifiability and revisability, ensuring a more interpretable and reliable reasoning process. Furthermore, we propose online revision by teacher models and offline revision with knowledge base. To enhance the relevance of knowledge retrieval from the knowledge base, we propose an antisense retrieval method to check if the newly generated knowledge contradicts any existing knowledge in the knowledge base to avoid retrieving meta-knowledge irrelevant to the problem. The experimental results on the Winogrande dataset have corroborated the efficacy of our proposed method. We revised the meta-knowledge of GPT-3.5 with GPT-4, which enhanced the accuracy from 68.11% to 73.64%, an improvement of 5.53%.
Paper Type: Long
Research Area: Question Answering
Research Area Keywords: Question Answering
Contribution Types: Model analysis & interpretability, NLP engineering experiment
Languages Studied: English
Submission Number: 5249
Loading