Editing the Mind of Giants: An In-Depth Exploration of Pitfalls of Knowledge Editing in Large Language Models

ACL ARR 2024 June Submission685 Authors

12 Jun 2024 (modified: 02 Jul 2024)ACL ARR 2024 June SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Abstract: Knowledge editing is a rising technique for efficiently updating factual knowledge in large language models (LLMs) with minimal alteration of parameters. However, recent studies have identified concerning side effects, such as knowledge distortion and the deterioration of general abilities, that have emerged after editing. This paper conducts a comprehensive study of these side effects, providing a unified view of the challenges associated with knowledge editing in LLMs. We discuss related work and summarize potential research directions to overcome these limitations. Our experiments highlight the limitations of current knowledge editing methods, emphasizing the need for deeper understanding of inner knowledge structures of LLMs and improved knowledge editing methods.
Paper Type: Long
Research Area: Machine Learning for NLP
Research Area Keywords: knowledge editing, large language model
Contribution Types: Surveys
Languages Studied: English
Submission Number: 685
Loading