Demystifying Language Model Forgetting with Low-rank Example Associations

Published: 11 Jun 2025, Last Modified: 10 Jul 2025ES-FoMo IIIEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Continual learning; Catastrophic forgetting;
TL;DR: We analyze, predict, and mitigate upstream example forgetting in LLM fine-tuning.
Abstract: Unintended forgetting of learned knowledge imposes an inherent risk on fine-tuning LLMs with new task data. In this paper, we adapt a classical framework for risk management to address forgetting by predicting their occurrence and perform targeted mitigation. Specifically, we show empirically that the associations between learned tasks and forgotten upstream examples are often well-approximated with low-rank matrices in diverse setups. Leveraging such empirical associations collected in the past, we predict forgetting of upstream examples when fine-tuning on unseen tasks with matrix completion. At the fine-tuning stage, we upweight examples predicted to be more forgotten for replay and demonstrate statistically significantly reduced forgetting over held-out, never-replayed upstream data. In summary, we show that predicting forgetting with low-rank example associations enables efficient examination and mitigation of forgetting.
Submission Number: 80
Loading