Abstract: Parameter-Efficient-Tuning (PET) for pre-trained deep models (e.g., transformer) hold significant potential for domain increment learning (DIL). Recent prevailing approaches resort to prompt learning, which typically involves learning a small number of prompts for each domain to avoid the issue of catastrophic forgetting. However, previous studies have pointed out prompt-based methods are often challenging to optimize, and their performance may vary non-monotonically with trainable parameters. In contrast to previous prompt-based DIL methods, we put forward an importance-aware shared parameter subspace learning for domain incremental learning, on the basis of low-rank adaption (LoRA). Specifically, we propose to incrementally learn a domain-specific and domain-shared low-rank parameter subspace for each domain, in order to effectively decouple the parameter space and capture shared information across different domains. Meanwhile, we present a momentum update strategy for learning the domain-shared subspace, allowing for the smoothly accumulation of knowledge in the current domain while mitigating the risk of forgetting the knowledge acquired from previous domains. Moreover, given that domain-shared information might hold varying degrees of importance across different domains, we design an importance-aware mechanism that adaptively assigns an importance weight to the domain-shared subspace for the corresponding domain. Finally, we devise a cross-domain contrastive constraint to encourage domain-specific subspaces to capture distinctive information within each domain effectively, and enforce orthogonality between domain-shared and domain-specific subspaces to minimize interference between them. Extensive experiments on image domain incremental datasets demonstrate the effectiveness of the proposed method in comparison to the related state-of-the-art methods.
Primary Subject Area: [Content] Multimodal Fusion
Relevance To Conference: Our method aims to propose a novel framework for handling data from multiple incremental domains. In fact, multiple domain data is commonly obtained from different sources and can be considered as data composed of multiple modalities. Therefore, this work will contribute to the exploration of multimodal processing in the continual learning field. Specifically, we propose a dynamic importance-aware parameter subspace decomposition for domain incremental learning, characterized by an flexible importance-conditioned domain-common and domain-specific subspace ensemble. We differentiate ourselves by considering the varying importance of subspaces across different domains. This research will provide insights into domain incremental learning of multimodal data and contribute to the advancement of multimedia and multimodal processing.
Supplementary Material: zip
Submission Number: 3826
Loading