LADA: Scalable Label-Specific CLIP Adapter for Continual Learning

Published: 01 May 2025, Last Modified: 18 Jun 2025ICML 2025 posterEveryoneRevisionsBibTeXCC BY 4.0
Abstract: Continual learning with vision-language models like CLIP offers a pathway toward scalable machine learning systems by leveraging its transferable representations. Existing CLIP-based methods adapt the pre-trained image encoder by adding multiple sets of learnable parameters, with each task using a partial set of parameters. This requires selecting the expected parameters for input images during inference, which is prone to error that degrades performance. To address this problem, we introduce LADA (**L**abel-specific **ADA**pter). Instead of partitioning parameters across tasks, LADA appends lightweight, label-specific memory units to the frozen CLIP image encoder, enabling discriminative feature generation by aggregating task-agnostic knowledge. To prevent catastrophic forgetting, LADA employs feature distillation for seen classes, preventing their features from being interfered with by new classes. Positioned after the image encoder, LADA prevents gradient flow to the frozen CLIP parameters, ensuring efficient training. Extensive results show that LADA achieves state-of-the-art performance in continual learning settings. The implementation code is available at [https://github.com/MaolinLuo/LADA](https://github.com/MaolinLuo/LADA).
Lay Summary: Continual learning enables AI systems to learn new tasks sequentially without forgetting previous knowledge. Existing methods for adapting vision-language models like CLIP require complex selection of task-specific parameters, which often leads to errors and degraded performance. We proposed LADA (Label-Specific Adapter), a simple yet powerful method that attaches compact, label-specific memory units to the fixed CLIP model. These units capture essential features from all tasks, allowing new tasks to be learned easily without interfering with prior knowledge. LADA significantly improves continual learning by eliminating the need for parameter selection, reducing errors, and ensuring strong performance across various tasks. Our method sets a new benchmark, helping AI models learn continually in a scalable and efficient way.
Link To Code: https://github.com/MaolinLuo/LADA
Primary Area: General Machine Learning->Supervised Learning
Keywords: Continual Learning, Vision-Language Model
Submission Number: 4500
Loading