ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.15699
14
0

BLUR: A Benchmark for LLM Unlearning Robust to Forget-Retain Overlap

28 May 2025
Shengyuan Hu
Neil Kale
Pratiksha Thaker
Yiwei Fu
Steven Wu
Virginia Smith
    MUAAMLCLL
ArXiv (abs)PDFHTML
Main:9 Pages
2 Figures
Bibliography:2 Pages
12 Tables
Appendix:3 Pages
Abstract

Machine unlearning has the potential to improve the safety of large language models (LLMs) by removing sensitive or harmful information post hoc. A key challenge in unlearning involves balancing between forget quality (effectively unlearning undesirable information) and retain quality (maintaining good performance on other, general tasks). Unfortunately, as we show, current LLM unlearning benchmarks contain highly disparate forget and retain sets -- painting a false picture of the effectiveness of LLM unlearning methods. This can be particularly problematic because it opens the door for benign perturbations, such as relearning attacks, to easily reveal supposedly unlearned knowledge once models are deployed. To address this, we present BLUR\texttt{BLUR}BLUR: a benchmark for LLM unlearning that provides more realistic scenarios of forget-retain overlap. BLUR\texttt{BLUR}BLUR significantly expands on existing unlearning benchmarks by providing extended evaluation tasks, combined forget/retain queries, and relearning datasets of varying degrees of difficulty. Despite the benign nature of the queries considered, we find that the performance of existing methods drops significantly when evaluated on BLUR\texttt{BLUR}BLUR, with simple approaches performing better on average than more recent methods. These results highlight the importance of robust evaluation and suggest several important directions of future study. Our benchmark is publicly available at:this https URL

View on arXiv
@article{hu2025_2506.15699,
  title={ BLUR: A Benchmark for LLM Unlearning Robust to Forget-Retain Overlap },
  author={ Shengyuan Hu and Neil Kale and Pratiksha Thaker and Yiwei Fu and Steven Wu and Virginia Smith },
  journal={arXiv preprint arXiv:2506.15699},
  year={ 2025 }
}
Comments on this paper