Lifelong Domain Adaptation via Consolidated Internal DistributionDownload PDF

21 May 2021, 20:42 (edited 26 Oct 2021)NeurIPS 2021 PosterReaders: Everyone
  • Keywords: lifelong machine learning, continual learning, unsupervised domain adaptation, internal distribution
  • TL;DR: In this paper, we propose an algorithm to update a model continually to tackle challenges of distributional shifts
  • Abstract: We develop an algorithm to address unsupervised domain adaptation (UDA) in continual learning (CL) settings. The goal is to update a model continually to learn distributional shifts across sequentially arriving tasks with unlabeled data while retaining the knowledge about the past learned tasks. Existing UDA algorithms address the challenge of domain shift, but they require simultaneous access to the datasets of the source and the target domains. On the other hand, existing works on CL can handle tasks with labeled data. Our solution is based on consolidating the learned internal distribution for improved model generalization on new domains and benefitting from experience replay to overcome catastrophic forgetting.
  • Supplementary Material: pdf
  • Code Of Conduct: I certify that all co-authors of this work have read and commit to adhering to the NeurIPS Statement on Ethics, Fairness, Inclusivity, and Code of Conduct.
  • Code: zip
25 Replies

Loading