The Twin Pillars of LLM Unlearning Audit: Ensuring Adequacy and Non-Redundancy through Knowledge Graphs

ACL ARR 2025 May Submission4697 Authors

20 May 2025 (modified: 03 Jul 2025)ACL ARR 2025 May SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Abstract: In recent years, LLMs have faced increasing demands to selectively remove sensitive information, protect privacy, and comply with copyright regulations through unlearning, by the Machine Unlearning. While evaluating unlearning effectiveness is crucial, existing benchmarks are limited in scale and comprehensiveness, typically containing only a few hundred test cases. We identify two critical challenges in generating holistic audit datasets: ensuring audit adequacy and handling knowledge redundancy between forget and retain dataset. To address these challenges, we propose HANKER, an automated framework for holistic audit dataset generation leveraging knowledge graphs to achieve fine-grained coverage and eliminate redundant knowledge. Applying HANKER to the popular MUSE benchmark, we successfully generated over 69,000 and 111,000 audit cases for the News and Books datasets respectively, identifying thousands of knowledge memorization instances that the previous benchmark failed to detect. Our empirical analysis uncovers how knowledge redundancy significantly skews unlearning effectiveness metrics, with redundant instances artificially inflating the observed memorization measurements ROUGE from 19.7% to 26.1% and Entailment Scores from 32.4% to 35.2%, highlighting the necessity of systematic deduplication for accurate assessment.
Paper Type: Long
Research Area: Language Modeling
Research Area Keywords: LLM Auditing, LLM Unlearning, Privacy Evaluation
Contribution Types: NLP engineering experiment, Data resources
Languages Studied: English
Submission Number: 4697
Loading