Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2105.00164
Cited By
v1
v2
v3 (latest)
Hidden Backdoors in Human-Centric Language Models
1 May 2021
Shaofeng Li
Hui Liu
Tian Dong
Benjamin Zi Hao Zhao
Minhui Xue
Haojin Zhu
Jialiang Lu
SILM
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Hidden Backdoors in Human-Centric Language Models"
29 / 79 papers shown
Title
Those Aren't Your Memories, They're Somebody Else's: Seeding Misinformation in Chat Bot Memories
Conor Atkins
Benjamin Zi Hao Zhao
Hassan Jameel Asghar
Ian D. Wood
M. Kâafar
KELM
36
2
0
06 Apr 2023
Do Backdoors Assist Membership Inference Attacks?
Yumeki Goto
Nami Ashizawa
Toshiki Shibahara
Naoto Yanai
MIACV
61
2
0
22 Mar 2023
NCL: Textual Backdoor Defense Using Noise-augmented Contrastive Learning
Shengfang Zhai
Qingni Shen
Xiaoyi Chen
Weilong Wang
Cong Li
Yuejian Fang
Zhonghai Wu
AAML
101
10
0
03 Mar 2023
FreeEagle: Detecting Complex Neural Trojans in Data-Free Cases
Chong Fu
Xuhong Zhang
S. Ji
Ting Wang
Peng Lin
Yanghe Feng
Jianwei Yin
AAML
70
11
0
28 Feb 2023
Attacks in Adversarial Machine Learning: A Systematic Survey from the Life-cycle Perspective
Baoyuan Wu
Zihao Zhu
Li Liu
Qingshan Liu
Zhaofeng He
Siwei Lyu
AAML
112
23
0
19 Feb 2023
Training-free Lexical Backdoor Attacks on Language Models
Yujin Huang
Terry Yue Zhuo
Xingliang Yuan
Han Hu
Lizhen Qu
Chunyang Chen
SILM
92
46
0
08 Feb 2023
BDMMT: Backdoor Sample Detection for Language Models through Model Mutation Testing
Jiali Wei
Ming Fan
Wenjing Jiao
Wuxia Jin
Ting Liu
AAML
99
15
0
25 Jan 2023
Stealthy Backdoor Attack for Code Models
Zhou Yang
Bowen Xu
Jie M. Zhang
Hong Jin Kang
Jieke Shi
Junda He
David Lo
AAML
60
68
0
06 Jan 2023
Mind Your Heart: Stealthy Backdoor Attack on Dynamic Deep Neural Network in Edge Computing
Tian Dong
Ziyuan Zhang
Han Qiu
Tianwei Zhang
Hewu Li
T. Wang
AAML
84
6
0
22 Dec 2022
Backdoor Vulnerabilities in Normally Trained Deep Learning Models
Guanhong Tao
Zhenting Wang
Shuyang Cheng
Shiqing Ma
Shengwei An
Yingqi Liu
Guangyu Shen
Zhuo Zhang
Yunshu Mao
Xiangyu Zhang
SILM
73
17
0
29 Nov 2022
On the Security Vulnerabilities of Text-to-SQL Models
Xutan Peng
Yipeng Zhang
Jingfeng Yang
Mark Stevenson
SILM
87
10
0
28 Nov 2022
BadPrompt: Backdoor Attacks on Continuous Prompts
Xiangrui Cai
Haidong Xu
Sihan Xu
Ying Zhang
Xiaojie Yuan
SILM
85
67
0
27 Nov 2022
A Survey on Backdoor Attack and Defense in Natural Language Processing
Xuan Sheng
Zhaoyang Han
Piji Li
Xiangmao Chang
SILM
71
21
0
22 Nov 2022
Poison Attack and Defense on Deep Source Code Processing Models
Jia Li
Zhuo Li
Huangzhao Zhang
Ge Li
Zhi Jin
Xing Hu
Xin Xia
AAML
69
19
0
31 Oct 2022
The "Beatrix'' Resurrections: Robust Backdoor Detection via Gram Matrices
Wanlun Ma
Derui Wang
Ruoxi Sun
Minhui Xue
S. Wen
Yang Xiang
AAML
94
86
0
23 Sep 2022
Kallima: A Clean-label Framework for Textual Backdoor Attacks
Xiaoyi Chen
Yinpeng Dong
Zeyu Sun
Shengfang Zhai
Qingni Shen
Zhonghai Wu
AAML
49
32
0
03 Jun 2022
MM-BD: Post-Training Detection of Backdoor Attacks with Arbitrary Backdoor Pattern Types Using a Maximum Margin Statistic
Hang Wang
Zhen Xiang
David J. Miller
G. Kesidis
AAML
96
44
0
13 May 2022
Exploring the Universal Vulnerability of Prompt-based Learning Paradigm
Lei Xu
Yangyi Chen
Ganqu Cui
Hongcheng Gao
Zhiyuan Liu
SILM
VPVLM
117
76
0
11 Apr 2022
PublicCheck: Public Integrity Verification for Services of Run-time Deep Models
Shuo Wang
Sharif Abuadbba
Sidharth Agarwal
Kristen Moore
Ruoxi Sun
Minhui Xue
Surya Nepal
S. Çamtepe
S. Kanhere
HILM
68
7
0
21 Mar 2022
Constrained Optimization with Dynamic Bound-scaling for Effective NLPBackdoor Defense
Guangyu Shen
Yingqi Liu
Guanhong Tao
Qiuling Xu
Zhuo Zhang
Shengwei An
Shiqing Ma
Xinming Zhang
AAML
81
41
0
11 Feb 2022
Spinning Language Models: Risks of Propaganda-As-A-Service and Countermeasures
Eugene Bagdasaryan
Vitaly Shmatikov
SILM
AAML
106
84
0
09 Dec 2021
Statically Detecting Adversarial Malware through Randomised Chaining
Matthew Crawford
Wei Wang
Ruoxi Sun
Minhui Xue
AAML
48
1
0
28 Nov 2021
Dissecting Malware in the Wild
H. Spencer
Wei Wang
Ruoxi Sun
Minhui Xue
25
1
0
28 Nov 2021
Data Hiding with Deep Learning: A Survey Unifying Digital Watermarking and Steganography
Zihan Wang
Olivia Byrnes
Hu Wang
Ruoxi Sun
Congbo Ma
Huaming Chen
Qi Wu
Minhui Xue
66
59
0
20 Jul 2021
Accumulative Poisoning Attacks on Real-time Data
Tianyu Pang
Xiao Yang
Yinpeng Dong
Hang Su
Jun Zhu
74
20
0
18 Jun 2021
Oriole: Thwarting Privacy against Trustworthy Deep Learning Models
Liuqiao Chen
Hu Wang
Benjamin Zi Hao Zhao
Minhui Xue
Hai-feng Qian
PICV
60
4
0
23 Feb 2021
Local and Central Differential Privacy for Robustness and Privacy in Federated Learning
Mohammad Naseri
Jamie Hayes
Emiliano De Cristofaro
FedML
124
151
0
08 Sep 2020
With Great Dispersion Comes Greater Resilience: Efficient Poisoning Attacks and Defenses for Linear Regression Models
Jialin Wen
Benjamin Zi Hao Zhao
Minhui Xue
Alina Oprea
Hai-feng Qian
AAML
70
20
0
21 Jun 2020
BadNL: Backdoor Attacks against NLP Models with Semantic-preserving Improvements
Xiaoyi Chen
A. Salem
Dingfan Chen
Michael Backes
Shiqing Ma
Qingni Shen
Zhonghai Wu
Yang Zhang
SILM
105
243
0
01 Jun 2020
Previous
1
2