Continual pre-training mitigates forgetting in language and vision

Published: 01 Jan 2024, Last Modified: 16 May 2025Neural Networks 2024EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: Highlights•Continual Pre-Training incrementally acquires knowledge from unstructured data streams.•Self-Supervised Continual Pre-Training effectively mitigates forgetting.•The representation drift is reduced by Self-Supervised Continual Pre-Training.•Performance on domain-specific tasks can be improved with a limited amount of data.
Loading