$\nabla \tau$: Gradient-based and Task-Agnostic Machine Unlearning

Published: 28 Jun 2024, Last Modified: 25 Jul 2024NextGenAISafety 2024 PosterEveryoneRevisionsBibTeXCC BY 4.0
Keywords: AI Safety, User Privacy, Machine Unlearning, Deep Learning, Membership Inference Attack
TL;DR: We present an efficient gradient-based framework for Machine Unlearning, capable of unlearning up to 30% of a dataset. Versatile across domains, our method improves performance by up to 10% over current methods without compromising accuracy.
Abstract: Machine Unlearning, the process of selectively eliminating the influence of certain data examples used during a model's training, is a crucial area of research for safeguarding User Privacy and ensuring compliance with recent data protection regulations. Existing unlearning methods face critical drawbacks, including their prohibitively high cost, often associated with a large number of hyperparameters, and the limitation of forgetting only relatively small data portions. This often makes retraining the model from scratch a quicker and more effective solution. In this study, we introduce Gradient-based and Task-Agnostic Machine Unlearning ($\nabla \tau$), an optimization framework designed to remove the influence of a subset of training data efficiently. It applies adaptive gradient ascent to the data to be forgotten while using standard gradient descent for the remaining data. $\nabla \tau$ offers multiple benefits over existing approaches. It enables the unlearning of large sections of the training dataset (up to 30%). It is versatile, supporting various unlearning tasks (such as subset forgetting or class removal) and applicable across different domains (images, text, etc.). Importantly, $\nabla \tau$ requires no hyperparameter adjustments, making it a more appealing option than retraining the model from scratch. We evaluate our framework's effectiveness using a set of well-established Membership Inference Attack metrics, demonstrating up to 10% enhancements in performance compared to state-of-the-art methods without compromising the original model's accuracy.
Submission Number: 137
Loading