Towards Efficient Machine Unlearning with Data Augmentation: Guided Loss-Increasing (GLI) to Prevent the Catastrophic Model Utility Drop
Abstract: Machine unlearning algorithms aim to make a model forget specific data that might be used in the training phase. To solve this problem, various studies have adopted loss-increasing methods. For example, some unlearning methods have presented data augmentation methods to generate synthesized images that maximize loss values for images to be forgotten. In contrast, some unlearning methods directly update the model in the direction of increasing loss for the images to be forgotten. In this paper, we first revisit these loss-increasing methods and analyze their limitations. We have found that these simple loss-increasing strategies can be effective in the aspect of the forgetting score, however, can hurt the original model utility unexpectedly, we call this phenomenon catastrophic model utility drop. We propose a novel data augmentation method, Guided Loss-Increasing (GLI), that restricts the direction of the data update to resolve the utility drop issue. This is achieved by aligning updates with the model’s existing knowledge, thereby ensuring that the unlearning process does not adversely affect the model’s original performance. Our extensive experiments demonstrate our method shows superior (1) model utility and (2) forgetting performance compared to the previous state-of-the-art (SOTA) methods. Furthermore, we demonstrate Jensen–Shannon divergence can be utilized to robustly evaluate the forgetting score. The source codes are publicly available at https://github.com/Dasol-Choi/Guided_Loss_Increasing.
Loading