Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2110.02467
Cited By
BadPre: Task-agnostic Backdoor Attacks to Pre-trained NLP Foundation Models
6 October 2021
Kangjie Chen
Yuxian Meng
Xiaofei Sun
Shangwei Guo
Tianwei Zhang
Jiwei Li
Chun Fan
SILM
Re-assign community
ArXiv
PDF
HTML
Papers citing
"BadPre: Task-agnostic Backdoor Attacks to Pre-trained NLP Foundation Models"
50 / 73 papers shown
Title
The Ripple Effect: On Unforeseen Complications of Backdoor Attacks
Rui Zhang
Yun Shen
Hongwei Li
Wenbo Jiang
Hanxiao Chen
Yuan Zhang
Guowen Xu
Yang Zhang
SILM
AAML
12
0
0
16 May 2025
BadLingual: A Novel Lingual-Backdoor Attack against Large Language Models
ziqi wang
Hongwei Li
Rui Zhang
Wenbo Jiang
Kangjie Chen
Tianwei Zhang
Qingchuan Zhao
Jiawei Li
AAML
46
0
0
06 May 2025
GaussTrap: Stealthy Poisoning Attacks on 3D Gaussian Splatting for Targeted Scene Confusion
Jiaxin Hong
Sixu Chen
Shuoyang Sun
Hongyao Yu
Hao Fang
Yuqi Tan
Bin Chen
Shuhan Qi
Jiawei Li
3DGS
AAML
135
0
0
29 Apr 2025
The Ultimate Cookbook for Invisible Poison: Crafting Subtle Clean-Label Text Backdoors with Style Attributes
Wencong You
Daniel Lowd
39
0
0
24 Apr 2025
SSD: A State-based Stealthy Backdoor Attack For Navigation System in UAV Route Planning
ziqi wang
Yang Li
Junzhe Zhang
Xingshuo Han
Kangbo Liu
Lyu Yang
yuan Zhou
Tianwei Zhang
Quan Pan
AAML
60
0
0
27 Feb 2025
Quantized Delta Weight Is Safety Keeper
Yule Liu
Zhen Sun
Xinlei He
Xinyi Huang
96
2
0
29 Nov 2024
RAG-Thief: Scalable Extraction of Private Data from Retrieval-Augmented Generation Applications with Agent-based Attacks
Changyue Jiang
Xudong Pan
Geng Hong
Chenfu Bao
Min Yang
SILM
75
9
0
21 Nov 2024
New Emerged Security and Privacy of Pre-trained Model: a Survey and Outlook
Meng Yang
Tianqing Zhu
Chi Liu
Wanlei Zhou
Shui Yu
Philip S. Yu
AAML
ELM
PILM
61
1
0
12 Nov 2024
CAT: Concept-level backdoor ATtacks for Concept Bottleneck Models
Songning Lai
Jiayu Yang
Yu Huang
Lijie Hu
Tianlang Xue
Zhangyi Hu
Jiaxu Li
Haicheng Liao
Yutao Yue
34
1
0
07 Oct 2024
Obliviate: Neutralizing Task-agnostic Backdoors within the Parameter-efficient Fine-tuning Paradigm
Jaehan Kim
Minkyoo Song
S. Na
Seungwon Shin
AAML
41
0
0
21 Sep 2024
The Dark Side of Human Feedback: Poisoning Large Language Models via User Inputs
Bocheng Chen
Hanqing Guo
Guangjing Wang
Yuanda Wang
Qiben Yan
AAML
37
4
0
01 Sep 2024
Rethinking Backdoor Detection Evaluation for Language Models
Jun Yan
Wenjie Jacky Mo
Xiang Ren
Robin Jia
ELM
51
1
0
31 Aug 2024
Turning Generative Models Degenerate: The Power of Data Poisoning Attacks
Shuli Jiang
S. Kadhe
Yi Zhou
Farhan Ahmed
Ling Cai
Nathalie Baracaldo
SILM
AAML
41
4
0
17 Jul 2024
Hey, That's My Model! Introducing Chain & Hash, An LLM Fingerprinting Technique
M. Russinovich
Ahmed Salem
51
12
0
15 Jul 2024
Distributed Backdoor Attacks on Federated Graph Learning and Certified Defenses
Yuxin Yang
Qiang Li
Jinyuan Jia
Yuan Hong
Binghui Wang
AAML
FedML
63
11
0
12 Jul 2024
Defending Code Language Models against Backdoor Attacks with Deceptive Cross-Entropy Loss
Guang Yang
Yu Zhou
Xiang Chen
Xiangyu Zhang
Terry Yue Zhuo
David Lo
Taolue Chen
AAML
57
4
0
12 Jul 2024
Unique Security and Privacy Threats of Large Language Model: A Comprehensive Survey
Shang Wang
Tianqing Zhu
Bo Liu
Ming Ding
Xu Guo
Dayong Ye
Wanlei Zhou
Philip S. Yu
PILM
67
17
0
12 Jun 2024
An LLM-Assisted Easy-to-Trigger Backdoor Attack on Code Completion Models: Injecting Disguised Vulnerabilities against Strong Detection
Shenao Yan
Shen Wang
Yue Duan
Hanbin Hong
Kiho Lee
Doowon Kim
Yuan Hong
AAML
SILM
43
17
0
10 Jun 2024
BadAgent: Inserting and Activating Backdoor Attacks in LLM Agents
Yifei Wang
Dizhan Xue
Shengjie Zhang
Shengsheng Qian
AAML
LLMAG
40
22
0
05 Jun 2024
Cross-Context Backdoor Attacks against Graph Prompt Learning
Xiaoting Lyu
Yufei Han
Wei Wang
Hangwei Qian
Ivor Tsang
Xiangliang Zhang
SILM
AAML
46
15
0
28 May 2024
TrojFM: Resource-efficient Backdoor Attacks against Very Large Foundation Models
Yuzhou Nie
Yanting Wang
Jinyuan Jia
Michael J. De Lucia
Nathaniel D. Bastian
Wenbo Guo
Dawn Song
SILM
AAML
36
5
0
27 May 2024
SEEP: Training Dynamics Grounds Latent Representation Search for Mitigating Backdoor Poisoning Attacks
Xuanli He
Qiongkai Xu
Jun Wang
Benjamin I. P. Rubinstein
Trevor Cohn
AAML
42
4
0
19 May 2024
BadEdit: Backdooring large language models by model editing
Yanzhou Li
Tianlin Li
Kangjie Chen
Jian Zhang
Shangqing Liu
Wenhan Wang
Tianwei Zhang
Yang Liu
SyDa
AAML
KELM
56
50
0
20 Mar 2024
WARDEN: Multi-Directional Backdoor Watermarks for Embedding-as-a-Service Copyright Protection
Anudeex Shetty
Yue Teng
Ke He
Qiongkai Xu
WaLM
30
5
0
03 Mar 2024
Double-I Watermark: Protecting Model Copyright for LLM Fine-tuning
Shen Li
Liuyi Yao
Jinyang Gao
Lan Zhang
Yaliang Li
49
11
0
22 Feb 2024
Purifying Large Language Models by Ensembling a Small Language Model
Tianlin Li
Qian Liu
Tianyu Pang
Chao Du
Qing Guo
Yang Liu
Min-Bin Lin
53
16
0
19 Feb 2024
Test-Time Backdoor Attacks on Multimodal Large Language Models
Dong Lu
Tianyu Pang
Chao Du
Qian Liu
Xianjun Yang
Min-Bin Lin
AAML
56
21
0
13 Feb 2024
OrderBkd: Textual backdoor attack through repositioning
Irina Alekseevskaia
Konstantin Arkhipenko
30
2
0
12 Feb 2024
Pre-trained Trojan Attacks for Visual Recognition
Aishan Liu
Xinwei Zhang
Yisong Xiao
Yuguang Zhou
Siyuan Liang
Jiakai Wang
Xianglong Liu
Xiaochun Cao
Dacheng Tao
AAML
71
25
0
23 Dec 2023
Forcing Generative Models to Degenerate Ones: The Power of Data Poisoning Attacks
Shuli Jiang
S. Kadhe
Yi Zhou
Ling Cai
Nathalie Baracaldo
SILM
AAML
17
13
0
07 Dec 2023
Foundation Models for Weather and Climate Data Understanding: A Comprehensive Survey
Shengchao Chen
Guodong Long
Jing Jiang
Dikai Liu
Chengqi Zhang
SyDa
AI4CE
41
24
0
05 Dec 2023
Grounding Foundation Models through Federated Transfer Learning: A General Framework
Yan Kang
Tao Fan
Hanlin Gu
Xiaojin Zhang
Lixin Fan
Qiang Yang
AI4CE
68
19
0
29 Nov 2023
Beyond Boundaries: A Comprehensive Survey of Transferable Attacks on AI Systems
Guangjing Wang
Ce Zhou
Yuanda Wang
Bocheng Chen
Hanqing Guo
Qiben Yan
AAML
SILM
66
3
0
20 Nov 2023
TextGuard: Provable Defense against Backdoor Attacks on Text Classification
Hengzhi Pei
Jinyuan Jia
Wenbo Guo
Bo-wen Li
Dawn Song
SILM
21
9
0
19 Nov 2023
Watermarking Vision-Language Pre-trained Models for Multi-modal Embedding as a Service
Yuanmin Tang
Jing Yu
Keke Gai
Xiangyang Qu
Yue Hu
Gang Xiong
Qi Wu
AAML
WaLM
VLM
24
7
0
10 Nov 2023
Last One Standing: A Comparative Analysis of Security and Privacy of Soft Prompt Tuning, LoRA, and In-Context Learning
Rui Wen
Tianhao Wang
Michael Backes
Yang Zhang
Ahmed Salem
AAML
21
10
0
17 Oct 2023
Privacy in Large Language Models: Attacks, Defenses and Future Directions
Haoran Li
Yulin Chen
Jinglong Luo
Yan Kang
Xiaojin Zhang
Qi Hu
Chunkit Chan
Yangqiu Song
PILM
48
42
0
16 Oct 2023
AFLOW: Developing Adversarial Examples under Extremely Noise-limited Settings
Renyang Liu
Jinhong Zhang
Haoran Li
Jin Zhang
Yuanyu Wang
Wei Zhou
AAML
30
3
0
15 Oct 2023
Composite Backdoor Attacks Against Large Language Models
Hai Huang
Zhengyu Zhao
Michael Backes
Yun Shen
Yang Zhang
AAML
32
36
0
11 Oct 2023
Fine-tuning Aligned Language Models Compromises Safety, Even When Users Do Not Intend To!
Xiangyu Qi
Yi Zeng
Tinghao Xie
Pin-Yu Chen
Ruoxi Jia
Prateek Mittal
Peter Henderson
SILM
70
525
0
05 Oct 2023
PETA: Parameter-Efficient Trojan Attacks
Lauren Hong
Ting Wang
AAML
46
1
0
01 Oct 2023
Backdoor Attacks and Countermeasures in Natural Language Processing Models: A Comprehensive Security Review
Pengzhou Cheng
Zongru Wu
Wei Du
Haodong Zhao
Wei Lu
Gongshen Liu
SILM
AAML
31
17
0
12 Sep 2023
A Comprehensive Overview of Backdoor Attacks in Large Language Models within Communication Networks
Haomiao Yang
Kunlan Xiang
Mengyu Ge
Hongwei Li
Rongxing Lu
Shui Yu
SILM
30
42
0
28 Aug 2023
LMSanitator: Defending Prompt-Tuning Against Task-Agnostic Backdoors
Chengkun Wei
Wenlong Meng
Zhikun Zhang
M. Chen
Ming-Hui Zhao
Wenjing Fang
Lei Wang
Zihui Zhang
Wenzhi Chen
AAML
29
8
0
26 Aug 2023
Use of LLMs for Illicit Purposes: Threats, Prevention Measures, and Vulnerabilities
Maximilian Mozes
Xuanli He
Bennett Kleinberg
Lewis D. Griffin
39
78
0
24 Aug 2023
GPTEval: A Survey on Assessments of ChatGPT and GPT-4
Rui Mao
Guanyi Chen
Xulang Zhang
Frank Guerin
Erik Cambria
ELM
LM&MA
36
101
0
24 Aug 2023
TIJO: Trigger Inversion with Joint Optimization for Defending Multimodal Backdoored Models
Indranil Sur
Karan Sikka
Matthew Walmer
K. Koneripalli
Anirban Roy
Xiaoyu Lin
Ajay Divakaran
Susmit Jha
32
8
0
07 Aug 2023
Towards Stealthy Backdoor Attacks against Speech Recognition via Elements of Sound
Hanbo Cai
Pengcheng Zhang
Hai Dong
Yan Xiao
Stefanos Koffas
Yiming Li
AAML
29
28
0
17 Jul 2023
Multi-target Backdoor Attacks for Code Pre-trained Models
Yanzhou Li
Shangqing Liu
Kangjie Chen
Xiaofei Xie
Tianwei Zhang
Yang Liu
AAML
SILM
20
23
0
14 Jun 2023
NOTABLE: Transferable Backdoor Attacks Against Prompt-based NLP Models
Kai Mei
Zheng Li
Zhenting Wang
Yang Zhang
Shiqing Ma
AAML
SILM
37
48
0
28 May 2023
1
2
Next