Lazy Safety Alignment for Large Language Models against Harmful
  Fine-tuning
v1v2v3v4 (latest)

Lazy Safety Alignment for Large Language Models against Harmful Fine-tuning

Sihao Hu
Fatih Ilhan
Selim Furkan Tekin
Ling Liu

Papers citing "Lazy Safety Alignment for Large Language Models against Harmful Fine-tuning"

28 / 28 papers shown
Title