Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2407.12281
Cited By
Turning Generative Models Degenerate: The Power of Data Poisoning Attacks
17 July 2024
Shuli Jiang
S. Kadhe
Yi Zhou
Farhan Ahmed
Ling Cai
Nathalie Baracaldo
SILM
AAML
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Turning Generative Models Degenerate: The Power of Data Poisoning Attacks"
7 / 7 papers shown
Title
Data Poisoning in Deep Learning: A Survey
Pinlong Zhao
Weiyao Zhu
Pengfei Jiao
Di Gao
Ou Wu
AAML
39
0
0
27 Mar 2025
Emergent Misalignment: Narrow finetuning can produce broadly misaligned LLMs
Jan Betley
Daniel Tan
Niels Warncke
Anna Sztyber-Betley
Xuchan Bao
Martín Soto
Nathan Labenz
Owain Evans
AAML
80
9
0
24 Feb 2025
Char-mander Use mBackdoor! A Study of Cross-lingual Backdoor Attacks in Multilingual LLMs
Himanshu Beniwal
Sailesh Panda
Birudugadda Srivibhav
Mayank Singh
45
0
0
24 Feb 2025
The Art of Deception: Robust Backdoor Attack using Dynamic Stacking of Triggers
Orson Mengara
AAML
49
3
0
03 Jan 2024
Prompt as Triggers for Backdoor Attack: Examining the Vulnerability in Language Models
Shuai Zhao
Jinming Wen
Anh Tuan Luu
J. Zhao
Jie Fu
SILM
62
89
0
02 May 2023
The Power of Scale for Parameter-Efficient Prompt Tuning
Brian Lester
Rami Al-Rfou
Noah Constant
VPVLM
280
3,848
0
18 Apr 2021
Text Summarization with Pretrained Encoders
Yang Liu
Mirella Lapata
MILM
258
1,432
0
22 Aug 2019
1