120

JPU: Bridging Jailbreak Defense and Unlearning via On-Policy Path Rectification

Xi Wang
Songlei Jian
Shasha Li
Xiaopeng Li
Zhaoye Li
Bin Ji
Baosheng Wang
Jie Yu
Main:7 Pages
6 Figures
Bibliography:3 Pages
3 Tables
Appendix:4 Pages
Abstract

Despite extensive safety alignment, Large Language Models (LLMs) often fail against jailbreak attacks. While machine unlearning has emerged as a promising defense by erasing specific harmful parameters, current methods remain vulnerable to diverse jailbreaks. We first conduct an empirical study and discover that this failure mechanism is caused by jailbreaks primarily activating non-erased parameters in the intermediate layers. Further, by probing the underlying mechanism through which these circumvented parameters reassemble into the prohibited output, we verify the persistent existence of dynamic jailbreak paths\textbf{jailbreak paths} and show that the inability to rectify them constitutes the fundamental gap in existing unlearning defenses. To bridge this gap, we propose J\textbf{J}ailbreak P\textbf{P}ath U\textbf{U}nlearning (JPU), which is the first to rectify dynamic jailbreak paths towards safety anchors by dynamically mining on-policy adversarial samples to expose vulnerabilities and identify jailbreak paths. Extensive experiments demonstrate that JPU significantly enhances jailbreak resistance against dynamic attacks while preserving the model's utility.

View on arXiv
Comments on this paper