Updating knowledge in Large Language Models: an Empirical Evaluation

Published: 01 Jan 2024, Last Modified: 08 Jan 2025EAIS 2024EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: Natural Language Processing (NLP) has witnessed a paradigm shift with Large Language Models (LLMs), yet the static knowledge from pre-training can lead to knowledge obsolescence. This study focuses on the dynamic relationship between LLMs and evolving knowledge, using GPT-2 as a case study. Leveraging an existing framework, we update models with monthly Wikipedia dumps and Wikidata probes, addressing the stability-plasticity trade-off. We introduce a novel synthetic data generation method for experimental control and present SMARTREVIEW, a state-of-the-art continual learning method. This work advances understanding and methodologies in tackling knowledge obsolescence in evolving language models.
Loading