Continual HyperTransformer: A Meta-Learner for Continual Few-Shot Learning

Published: 03 May 2024, Last Modified: 03 May 2024Accepted by TMLREveryoneRevisionsBibTeX
Abstract: We focus on the problem of learning without forgetting from multiple tasks arriving sequentially, where each task is defined using a few-shot episode of novel or already seen classes. We approach this problem using the recently published HyperTransformer (HT), a Transformer-based hypernetwork that generates specialized task-specific CNN weights directly from the support set. In order to learn from a continual sequence of tasks, we propose to recursively re-use the generated weights as input to the HT for the next task. This way, the generated CNN weights themselves act as a representation of previously learned tasks, and the HT is trained to update these weights so that the new task can be learned without forgetting past tasks. This approach is different from most continual learning algorithms that typically rely on using replay buffers, weight regularization or task-dependent architectural changes. We demonstrate that our proposed Continual HyperTransformer method equipped with a prototypical loss is capable of learning and retaining knowledge about past tasks for a variety of scenarios, including learning from mini-batches, and task-incremental and class-incremental learning scenarios.
Submission Length: Regular submission (no more than 12 pages of main content)
Changes Since Last Submission: Added details regarding the HyperTransformer and the Continual HyperTransformer architectures. Added a comment about the confidence intervals to Figures 3 and 4.
Assigned Action Editor: ~Jake_Snell1
Submission Number: 1631