Retaining Knowledge for Learning with Dynamic DefinitionDownload PDF

Published: 31 Oct 2022, 18:00, Last Modified: 16 Jan 2023, 04:30NeurIPS 2022 AcceptReaders: Everyone
TL;DR: We propose a new model architecture for use in a dynamic learning setting that frequently appears in practice. We prove (theoretically and empirically) that our method outperforms existing techniques.
Abstract: Machine learning models are often deployed in settings where they must be constantly updated in response to the changes in class definitions while retaining high accuracy on previously learned definitions. A classical use case is fraud detection, where new fraud schemes come one after another. While such an update can be accomplished by re-training on the complete data, the process is inefficient and prevents real-time and on-device learning. On the other hand, efficient methods that incrementally learn from new data often result in the forgetting of previously-learned knowledge. We define this problem as Learning with Dynamic Definition (LDD) and demonstrate that popular models, such as the Vision Transformer and Roberta, exhibit substantial forgetting of past definitions. We present the first practical and provable solution to LDD. Our proposal is a hash-based sparsity model \textit{RIDDLE} that solves evolving definitions by associating samples only to relevant parameters. We prove that our model is a universal function approximator and theoretically bounds the knowledge lost during the update process. On practical tasks with evolving class definition in vision and natural language processing, \textit{RIDDLE} outperforms baselines by up to 30\% on the original dataset while providing competitive accuracy on the update dataset.
Supplementary Material: zip
16 Replies

Loading