Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1812.05271
Cited By
TextBugger: Generating Adversarial Text Against Real-world Applications
13 December 2018
Jinfeng Li
S. Ji
Tianyu Du
Bo Li
Ting Wang
SILM
AAML
Re-assign community
ArXiv
PDF
HTML
Papers citing
"TextBugger: Generating Adversarial Text Against Real-world Applications"
50 / 382 papers shown
Title
Hijacking Large Language Models via Adversarial In-Context Learning
Yao Qiang
Xiangyu Zhou
Dongxiao Zhu
32
32
0
16 Nov 2023
Whispers of Doubt Amidst Echoes of Triumph in NLP Robustness
Ashim Gupta
Rishanth Rajendhran
Nathan Stringham
Vivek Srikumar
Ana Marasović
AAML
31
3
0
16 Nov 2023
DALA: A Distribution-Aware LoRA-Based Adversarial Attack against Language Models
Yibo Wang
Xiangjue Dong
James Caverlee
Philip S. Yu
29
2
0
14 Nov 2023
Robust Text Classification: Analyzing Prototype-Based Networks
Zhivar Sourati
D. Deshpande
Filip Ilievski
Kiril Gashteovski
S. Saralajew
OOD
OffRL
39
2
0
11 Nov 2023
Towards Effective Paraphrasing for Information Disguise
Anmol Agarwal
Shrey Gupta
Vamshi Krishna Bonagiri
Manas Gaur
Joseph M. Reagle
Ponnurangam Kumaraguru
40
3
0
08 Nov 2023
Robustifying Language Models with Test-Time Adaptation
Noah T. McDermott
Junfeng Yang
Chengzhi Mao
24
2
0
29 Oct 2023
BERT Lost Patience Won't Be Robust to Adversarial Slowdown
Zachary Coalson
Gabriel Ritter
Rakesh Bobba
Sanghyun Hong
AAML
24
1
0
29 Oct 2023
CT-GAT: Cross-Task Generative Adversarial Attack based on Transferability
Minxuan Lv
Chengwei Dai
Kun Li
Wei Zhou
Song Hu
AAML
40
6
0
22 Oct 2023
Toward Stronger Textual Attack Detectors
Pierre Colombo
Marine Picot
Nathan Noiry
Guillaume Staerman
Pablo Piantanida
59
5
0
21 Oct 2023
An LLM can Fool Itself: A Prompt-Based Adversarial Attack
Xilie Xu
Keyi Kong
Ning Liu
Li-zhen Cui
Di Wang
Jingfeng Zhang
Mohan Kankanhalli
AAML
SILM
36
68
0
20 Oct 2023
Towards Robust Pruning: An Adaptive Knowledge-Retention Pruning Strategy for Language Models
Jianwei Li
Qi Lei
Wei Cheng
Dongkuan Xu
KELM
33
3
0
19 Oct 2023
To Generate or Not? Safety-Driven Unlearned Diffusion Models Are Still Easy To Generate Unsafe Images ... For Now
Yimeng Zhang
Jinghan Jia
Xin Chen
Aochuan Chen
Yihua Zhang
Jiancheng Liu
Ke Ding
Sijia Liu
DiffM
22
83
0
18 Oct 2023
Survey of Vulnerabilities in Large Language Models Revealed by Adversarial Attacks
Erfan Shayegani
Md Abdullah Al Mamun
Yu Fu
Pedram Zaree
Yue Dong
Nael B. Abu-Ghazaleh
AAML
147
148
0
16 Oct 2023
BufferSearch: Generating Black-Box Adversarial Texts With Lower Queries
Wenjie Lv
Zhen Wang
Yitao Zheng
Zhehua Zhong
Qi Xuan
Tianyi Chen
AAML
33
0
0
14 Oct 2023
VLATTACK: Multimodal Adversarial Attacks on Vision-Language Tasks via Pre-trained Models
Ziyi Yin
Muchao Ye
Tianrong Zhang
Tianyu Du
Jinguo Zhu
Han Liu
Jinghui Chen
Ting Wang
Fenglong Ma
AAML
VLM
CoGe
33
36
0
07 Oct 2023
SmoothLLM: Defending Large Language Models Against Jailbreaking Attacks
Alexander Robey
Eric Wong
Hamed Hassani
George J. Pappas
AAML
49
220
0
05 Oct 2023
Fooling the Textual Fooler via Randomizing Latent Representations
Duy C. Hoang
Quang H. Nguyen
Saurav Manchanda
MinLong Peng
Kok-Seng Wong
Khoa D. Doan
SILM
AAML
23
0
0
02 Oct 2023
DyVal: Dynamic Evaluation of Large Language Models for Reasoning Tasks
A. Maritan
Jiaao Chen
S. Dey
Luca Schenato
Diyi Yang
Xing Xie
ELM
LRM
27
42
0
29 Sep 2023
Defending Against Alignment-Breaking Attacks via Robustly Aligned LLM
Bochuan Cao
Yu Cao
Lu Lin
Jinghui Chen
AAML
36
135
0
18 Sep 2023
AGent: A Novel Pipeline for Automatically Creating Unanswerable Questions
Son Quoc Tran
Gia-Huy Do
Phong Nguyen-Thuan Do
Matt Kretchmar
Xinya Du
29
0
0
10 Sep 2023
Learning from Limited Heterogeneous Training Data: Meta-Learning for Unsupervised Zero-Day Web Attack Detection across Web Domains
Peiyang Li
Ye Wang
Qi Li
Zhuotao Liu
Ke Xu
Ju Ren
Zhiying Liu
Ruilin Lin
AAML
29
8
0
07 Sep 2023
Efficient Query-Based Attack against ML-Based Android Malware Detection under Zero Knowledge Setting
Ping He
Yifan Xia
Xuhong Zhang
Shouling Ji
AAML
20
12
0
05 Sep 2023
MathAttack: Attacking Large Language Models Towards Math Solving Ability
Zihao Zhou
Qiufeng Wang
Mingyu Jin
Jie Yao
Jianan Ye
Wei Liu
Wei Wang
Xiaowei Huang
Kaizhu Huang
AAML
32
26
0
04 Sep 2023
Baseline Defenses for Adversarial Attacks Against Aligned Language Models
Neel Jain
Avi Schwarzschild
Yuxin Wen
Gowthami Somepalli
John Kirchenbauer
Ping Yeh-Chiang
Micah Goldblum
Aniruddha Saha
Jonas Geiping
Tom Goldstein
AAML
60
340
0
01 Sep 2023
Vulnerability of Machine Learning Approaches Applied in IoT-based Smart Grid: A Review
Zhenyong Zhang
Mengxiang Liu
Mingyang Sun
Ruilong Deng
Peng Cheng
Dusit Niyato
Mo-Yuen Chow
Jiming Chen
40
39
0
30 Aug 2023
Use of LLMs for Illicit Purposes: Threats, Prevention Measures, and Vulnerabilities
Maximilian Mozes
Xuanli He
Bennett Kleinberg
Lewis D. Griffin
39
78
0
24 Aug 2023
LEAP: Efficient and Automated Test Method for NLP Software
Ming-Ming Xiao
Yan Xiao
Hai Dong
Shunhui Ji
Pengcheng Zhang
AAML
24
8
0
22 Aug 2023
An Image is Worth a Thousand Toxic Words: A Metamorphic Testing Framework for Content Moderation Software
Wenxuan Wang
Jingyuan Huang
Jen-tse Huang
Chang Chen
Jiazhen Gu
Pinjia He
Michael R. Lyu
VLM
36
6
0
18 Aug 2023
Robustness Over Time: Understanding Adversarial Examples' Effectiveness on Longitudinal Versions of Large Language Models
Yugeng Liu
Tianshuo Cong
Zhengyu Zhao
Michael Backes
Yun Shen
Yang Zhang
AAML
41
6
0
15 Aug 2023
Automated Testing and Improvement of Named Entity Recognition Systems
Boxi Yu
Yi-Nuo Hu
Qiuyang Mang
Wen-Ying Hu
Pinjia He
25
6
0
14 Aug 2023
Mondrian: Prompt Abstraction Attack Against Large Language Models for Cheaper API Pricing
Waiman Si
Michael Backes
Yang Zhang
30
5
0
07 Aug 2023
URET: Universal Robustness Evaluation Toolkit (for Evasion)
Kevin Eykholt
Taesung Lee
D. Schales
Jiyong Jang
Ian Molloy
Masha Zorin
AAML
41
7
0
03 Aug 2023
LimeAttack: Local Explainable Method for Textual Hard-Label Adversarial Attack
HaiXiang Zhu
Zhaoqing Yang
Weiwei Shang
Yuren Wu
AAML
FAtt
10
3
0
01 Aug 2023
Text-CRS: A Generalized Certified Robustness Framework against Textual Adversarial Attacks
Xinyu Zhang
Hanbin Hong
Yuan Hong
Peng Huang
Binghui Wang
Zhongjie Ba
Kui Ren
SILM
42
18
0
31 Jul 2023
Lost In Translation: Generating Adversarial Examples Robust to Round-Trip Translation
Neel Bhandari
Pin-Yu Chen
AAML
SILM
45
3
0
24 Jul 2023
FATRER: Full-Attention Topic Regularizer for Accurate and Robust Conversational Emotion Recognition
Yuzhao Mao
Di Lu
Xiaojie Wang
Yang Zhang
27
1
0
23 Jul 2023
Making Pre-trained Language Models both Task-solvers and Self-calibrators
Yangyi Chen
Xingyao Wang
Heng Ji
20
0
0
21 Jul 2023
Certified Robustness for Large Language Models with Self-Denoising
Zhen Zhang
Guanhua Zhang
Bairu Hou
Wenqi Fan
Qing Li
Sijia Liu
Yang Zhang
Shiyu Chang
19
22
0
14 Jul 2023
Interpretability and Transparency-Driven Detection and Transformation of Textual Adversarial Examples (IT-DT)
Bushra Sabir
Muhammad Ali Babar
Sharif Abuadbba
SILM
42
8
0
03 Jul 2023
Shilling Black-box Review-based Recommender Systems through Fake Review Generation
Hung-Yun Chiang
Yi-Syuan Chen
Yun-Zhu Song
Hong-Han Shuai
Jason J. S. Chang
AAML
13
12
0
27 Jun 2023
A Survey on Out-of-Distribution Evaluation of Neural NLP Models
Xinzhe Li
Ming Liu
Shang Gao
Wray L. Buntine
14
20
0
27 Jun 2023
DSRM: Boost Textual Adversarial Training with Distribution Shift Risk Minimization
Songyang Gao
Shihan Dou
Yan Liu
Xiao Wang
Qi Zhang
Zhongyu Wei
Jin Ma
Yingchun Shan
OOD
25
3
0
27 Jun 2023
Sample Attackability in Natural Language Adversarial Attacks
Vyas Raina
Mark Gales
SILM
45
1
0
21 Jun 2023
Investigating Masking-based Data Generation in Language Models
Edward Ma
35
0
0
16 Jun 2023
Evaluating the Robustness of Text-to-image Diffusion Models against Real-world Attacks
Hongcheng Gao
Hao Zhang
Yinpeng Dong
Zhijie Deng
AAML
41
21
0
16 Jun 2023
Explore, Establish, Exploit: Red Teaming Language Models from Scratch
Stephen Casper
Jason Lin
Joe Kwon
Gatlen Culp
Dylan Hadfield-Menell
AAML
8
83
0
15 Jun 2023
When Vision Fails: Text Attacks Against ViT and OCR
Nicholas Boucher
Jenny Blessing
Ilia Shumailov
Ross J. Anderson
Nicolas Papernot
AAML
42
4
0
12 Jun 2023
Enhancing Robustness of AI Offensive Code Generators via Data Augmentation
Cristina Improta
Pietro Liguori
R. Natella
B. Cukic
Domenico Cotroneo
AAML
32
2
0
08 Jun 2023
Expanding Scope: Adapting English Adversarial Attacks to Chinese
Hanyu Liu
Chengyuan Cai
Yanjun Qi
AAML
29
5
0
08 Jun 2023
Revisiting Out-of-distribution Robustness in NLP: Benchmark, Analysis, and LLMs Evaluations
Lifan Yuan
Yangyi Chen
Ganqu Cui
Hongcheng Gao
Fangyuan Zou
Xingyi Cheng
Heng Ji
Zhiyuan Liu
Maosong Sun
39
73
0
07 Jun 2023
Previous
1
2
3
4
5
6
7
8
Next