Data-Unlearn-Bench: Making Evaluating Data Unlearning Easy

Published: 11 Jun 2025, Last Modified: 11 Jun 2025MUGen @ ICML 2025 PosterEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Machine Unlearning, Benchmark, KL divergence of margins, Reproducibility
TL;DR: Data-Unlearn-Bench simplifies machine unlearning evaluation using the KL divergence of margins (KLoM) metric with precomputed pre-trained models and oracles, enabling reproducible, low-overhead, and fair comparisons to accelerate research.
Abstract:

Evaluating machine unlearning methods remains technically challenging, with recent benchmarks requiring complex setups and significant engineering overhead. We introduce a unified and extensible benchmarking suite that simplifies the evaluation of unlearning algorithms using the KLoM (KL divergence of Margins) metric. Our framework provides precomputed model ensembles, oracle outputs, and streamlined infrastructure for running evaluations out of the box. By standardizing setup and metrics, it enables reproducible, scalable, and fair comparison across unlearning methods. We aim for this benchmark to serve as a practical foundation for accelerating research and promoting best practices in machine unlearning. Our code and data are publicly available.

Submission Number: 52
Loading