ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2407.00106
  4. Cited By
UnUnlearning: Unlearning is not sufficient for content regulation in
  advanced generative AI

UnUnlearning: Unlearning is not sufficient for content regulation in advanced generative AI

27 June 2024
Ilia Shumailov
Jamie Hayes
Eleni Triantafillou
Guillermo Ortiz-Jimenez
Nicolas Papernot
Matthew Jagielski
Itay Yona
Heidi Howard
Eugene Bagdasaryan
    MU
ArXiv (abs)PDFHTML

Papers citing "UnUnlearning: Unlearning is not sufficient for content regulation in advanced generative AI"

8 / 8 papers shown
Title
Rectifying Privacy and Efficacy Measurements in Machine Unlearning: A New Inference Attack Perspective
Rectifying Privacy and Efficacy Measurements in Machine Unlearning: A New Inference Attack Perspective
Nima Naderloui
Shenao Yan
Binghui Wang
Jie Fu
Wendy Hui Wang
Weiran Liu
Yuan Hong
AAML
37
0
0
16 Jun 2025
Prompt Attacks Reveal Superficial Knowledge Removal in Unlearning Methods
Prompt Attacks Reveal Superficial Knowledge Removal in Unlearning Methods
Yeonwoo Jang
Shariqah Hossain
Ashwin Sreevatsa
Diogo Cruz
AAMLMU
54
0
0
11 Jun 2025
LLM Unlearning Should Be Form-Independent
LLM Unlearning Should Be Form-Independent
Xiaotian Ye
Mengqi Zhang
Shu Wu
MU
27
0
0
09 Jun 2025
Representation Bending for Large Language Model Safety
Representation Bending for Large Language Model Safety
Ashkan Yousefpour
Taeheon Kim
Ryan S. Kwon
Seungbeen Lee
Wonje Jeung
Seungju Han
Alvin Wan
Harrison Ngan
Youngjae Yu
Jonghyun Choi
AAMLALMKELM
129
4
0
02 Apr 2025
Adversarial ML Problems Are Getting Harder to Solve and to Evaluate
Adversarial ML Problems Are Getting Harder to Solve and to Evaluate
Javier Rando
Jie Zhang
Nicholas Carlini
F. Tramèr
AAMLELM
139
9
0
04 Feb 2025
Model Tampering Attacks Enable More Rigorous Evaluations of LLM Capabilities
Model Tampering Attacks Enable More Rigorous Evaluations of LLM Capabilities
Zora Che
Stephen Casper
Robert Kirk
Anirudh Satheesh
Stewart Slocum
...
Zikui Cai
Bilal Chughtai
Y. Gal
Furong Huang
Dylan Hadfield-Menell
MUAAMLELM
181
7
0
03 Feb 2025
Mitigating Memorization In Language Models
Mitigating Memorization In Language Models
Mansi Sakarvadia
Aswathy Ajith
Arham Khan
Nathaniel Hudson
Caleb Geniesse
Kyle Chard
Yaoqing Yang
Ian Foster
Michael W. Mahoney
KELMMU
130
2
0
03 Oct 2024
An Adversarial Perspective on Machine Unlearning for AI Safety
An Adversarial Perspective on Machine Unlearning for AI Safety
Jakub Łucki
Boyi Wei
Yangsibo Huang
Peter Henderson
F. Tramèr
Javier Rando
MUAAML
206
53
0
26 Sep 2024
1