CD-IMM: The Benefits of Domain-based Mixture Models in Bayesian Continual Learning

Published: 15 Aug 2024, Last Modified: 15 Aug 20241st CLAI Unconf Stage 2EveryoneRevisionsBibTeXCC BY 4.0
Keywords: continual learning, bayesian models
Abstract: Real-world streams of data are characterised by the continuous occurrence of new and old classes, possibly on novel domains. Bayesian non-parametric mixture models provide a natural solution for continual learning due to their ability to create new components on the fly when new data are observed. However, popular class-based and time-based mixtures are often tested on simplified streams (\eg class-incremental), where shortcuts can be exploited to infer drifts. We hypothesise that \emph{domain-based mixtures are more effective on natural streams}. Our proposed method, the CD-IMM, exemplifies this approach by learning an infinite mixture of domains for each class. We experiment on a natural scenario with a mix of class repetitions and novel domains to validate our hypothesis. The experimental results confirm our hypothesis and we find that CD-IMM beats state-of-the-art bayesian continual learning methods.
Submission Number: 3
Loading