Abstract: Continual learning (CL) refers to the ability to continuously learn and accumulate new knowledge while retaining useful information from past experiences. Although numerous CL methods have been proposed in recent years, it is not straightforward to deploy them directly to real-world decision-making problems due to their computational cost and lack of uncertainty quantification. To address these issues, we propose CL-BRUNO, a probabilistic, Neural Process-based CL model that performs scalable and tractable Bayesian update and prediction via probabilistic exchangeable sequence modelling. Our proposed approach uses deep-generative models to create a unified Bayesian probabilistic framework capable of handling different types of CL problems such as task- and class-incremental learning by modelling data from different tasks as sequences of exchangeable random variables, allowing users to integrate information across different CL scenarios efficiently using a single model, and give easy-to-interpret probabilistic predictions without the need of training or maintaining separate classifiers. Our approach is able to prevent catastrophic forgetting through distributional and functional regularisation without the need of retaining any previously seen samples, making it appealing to applications where data privacy or storage capacity is of concern. Experiments show that CL-BRUNO outperforms existing methods on both natural image and biomedical data sets, confirming its effectiveness in real-world applications.
Submission Length: Long submission (more than 12 pages of main content)
Changes Since Last Submission: We fixed typos, revised the benchmarking section, incorporated additional ablation studies, rearranged figures and tables, and added a Broader Impact Statement. We also revised text and corrected typos. We also added an algorithm box summarising the proposed model.
Assigned Action Editor: ~Magda_Gregorova2
Submission Number: 4714
Loading