Q-Tuning: Continual Queue-based Prompt Tuning for Language Models

15 Sept 2023 (modified: 25 Mar 2024)ICLR 2024 Conference Withdrawn SubmissionEveryoneRevisionsBibTeX
Keywords: Continual Leanring, Prompt Tuning, Continual Prompt Tuning, Language Model
TL;DR: This paper introduces a continual prompt tuning method called Q-tuning that, to our knowledge, is the first technique for achieving lifelong learning on extremely long task sequences through prompt tuning.
Abstract: This paper introduces **Q-tuning**, a novel approach for continual prompt tuning that enables the lifelong learning of a pretrained language model on a sequence of tasks. For each new task, Q-tuning trains a task-specific prompt by adding it to the prompt queue consisting of the prompts from older tasks. To better transfer the knowledge of older tasks, we design an ensemble mechanism that reweighs previous prompts in queue with a learnable low-rank matrix that reflects their relevance to the current task. To facilitate training and inference with manageable complexity, once the prompt queue reaches its maximum capacity, we leverage a PCA-based eviction rule to reduce the queue's size, allowing the newly trained prompt to be added while preserving the primary knowledge of older tasks. In order to mitigate the accumulation of information loss caused by the eviction, we additionally propose a globally shared prefix prompt and a memory retention regularization based on the information theory. Extensive experiments demonstrate that our approach outperforms the state-of-the-art methods substantially on both short and long task sequences. Moreover, our approach enables the lifelong learning on an extremely long task sequence while requiring only $\mathcal{O}(1)$ complexity for training and inference, which could not be achieved by existing technologies.
Primary Area: transfer learning, meta learning, and lifelong learning
Code Of Ethics: I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics.
Submission Guidelines: I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2024/AuthorGuide.
Anonymous Url: I certify that there is no URL (e.g., github page) that could be used to find authors' identity.
No Acknowledgement Section: I certify that there is no acknowledgement section in this submission for double blind review.
Submission Number: 383
Loading