Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1812.05271
Cited By
TextBugger: Generating Adversarial Text Against Real-world Applications
13 December 2018
Jinfeng Li
S. Ji
Tianyu Du
Bo Li
Ting Wang
SILM
AAML
Re-assign community
ArXiv
PDF
HTML
Papers citing
"TextBugger: Generating Adversarial Text Against Real-world Applications"
50 / 382 papers shown
Title
Defensive Unlearning with Adversarial Training for Robust Concept Erasure in Diffusion Models
Yimeng Zhang
Xin Chen
Jinghan Jia
Yihua Zhang
Chongyu Fan
Jiancheng Liu
Mingyi Hong
Ke Ding
Sijia Liu
DiffM
44
53
0
24 May 2024
Unveiling the Achilles' Heel of NLG Evaluators: A Unified Adversarial Framework Driven by Large Language Models
Yiming Chen
Chen Zhang
Danqing Luo
L. F. D’Haro
R. Tan
Haizhou Li
AAML
ELM
40
2
0
23 May 2024
AI-Cybersecurity Education Through Designing AI-based Cyberharassment Detection Lab
Ebuka Okpala
Nishant Vishwamitra
Keyan Guo
Song Liao
Long Cheng
Hongxin Hu
Yongkai Wu
Xiaohong Yuan
Jeannette Wade
S. Khorsandroo
9
0
0
13 May 2024
Revisiting character-level adversarial attacks
Elias Abad Rocamora
Yongtao Wu
Fanghui Liu
Grigorios G. Chrysos
V. Cevher
AAML
39
3
0
07 May 2024
Assessing Adversarial Robustness of Large Language Models: An Empirical Study
Zeyu Yang
Zhao Meng
Xiaochen Zheng
Roger Wattenhofer
ELM
AAML
31
7
0
04 May 2024
Adversarial Attacks and Defense for Conversation Entailment Task
Zhenning Yang
Ryan Krawec
Liang-Yuan Wu
AAML
SILM
27
1
0
01 May 2024
Revisiting the Adversarial Robustness of Vision Language Models: a Multimodal Perspective
Wanqi Zhou
Shuanghao Bai
Qibin Zhao
Badong Chen
VLM
AAML
44
5
0
30 Apr 2024
GRAMMAR: Grounded and Modular Methodology for Assessment of Closed-Domain Retrieval-Augmented Language Model
Xinzhe Li
Ming Liu
Shang Gao
RALM
48
0
0
30 Apr 2024
Advancing the Robustness of Large Language Models through Self-Denoised Smoothing
Jiabao Ji
Bairu Hou
Zhen Zhang
Guanhua Zhang
Wenqi Fan
Qing Li
Yang Zhang
Gaowen Liu
Sijia Liu
Shiyu Chang
AAML
43
6
0
18 Apr 2024
Enhance Robustness of Language Models Against Variation Attack through Graph Integration
Ziteng Xiong
Lizhi Qing
Yangyang Kang
Jiawei Liu
Hongsong Li
Changlong Sun
Xiaozhong Liu
Wei Lu
26
1
0
18 Apr 2024
GenFighter: A Generative and Evolutive Textual Attack Removal
Md Athikul Islam
Edoardo Serra
Sushil Jajodia
AAML
29
0
0
17 Apr 2024
Resilience of Large Language Models for Noisy Instructions
Bin Wang
Chengwei Wei
Zhengyuan Liu
Geyu Lin
Nancy F. Chen
49
11
0
15 Apr 2024
VertAttack: Taking advantage of Text Classifiers' horizontal vision
Jonathan Rusert
AAML
43
1
0
12 Apr 2024
Towards Building a Robust Toxicity Predictor
Dmitriy Bespalov
Sourav S. Bhabesh
Yi Xiang
Liutong Zhou
Yanjun Qi
AAML
116
10
0
09 Apr 2024
Semantic Stealth: Adversarial Text Attacks on NLP Using Several Methods
Roopkatha Dey
Aivy Debnath
Sayak Kumar Dutta
Kaustav Ghosh
Arijit Mitra
Arghya Roy Chowdhury
Jaydip Sen
AAML
SILM
29
1
0
08 Apr 2024
Goal-guided Generative Prompt Injection Attack on Large Language Models
Chong Zhang
Mingyu Jin
Qinkai Yu
Chengzhi Liu
Haochen Xue
Xiaobo Jin
AAML
SILM
42
12
0
06 Apr 2024
READ: Improving Relation Extraction from an ADversarial Perspective
Dawei Li
William Hogan
Jingbo Shang
AAML
36
0
0
02 Apr 2024
Multi-granular Adversarial Attacks against Black-box Neural Ranking Models
Yuansan Liu
Ruqing Zhang
J. Guo
Maarten de Rijke
Yixing Fan
Xueqi Cheng
AAML
54
13
0
02 Apr 2024
PID Control-Based Self-Healing to Improve the Robustness of Large Language Models
Zhuotong Chen
Zihu Wang
Yifan Yang
Qianxiao Li
Zheng Zhang
AAML
42
1
0
31 Mar 2024
SemRoDe: Macro Adversarial Training to Learn Representations That are Robust to Word-Level Attacks
Brian Formento
Wenjie Feng
Chuan-Sheng Foo
Anh Tuan Luu
See-Kiong Ng
AAML
34
7
0
27 Mar 2024
Targeted Visualization of the Backbone of Encoder LLMs
Isaac Roberts
Alexander Schulz
L. Hermes
Barbara Hammer
45
0
0
26 Mar 2024
Subspace Defense: Discarding Adversarial Perturbations by Learning a Subspace for Clean Signals
Rui Zheng
Yuhao Zhou
Zhiheng Xi
Tao Gui
Qi Zhang
Xuanjing Huang
AAML
55
0
0
24 Mar 2024
Don't be a Fool: Pooling Strategies in Offensive Language Detection from User-Intended Adversarial Attacks
Seunguk Yu
Juhwan Choi
Youngbin Kim
AAML
21
0
0
20 Mar 2024
SSCAE -- Semantic, Syntactic, and Context-aware natural language Adversarial Examples generator
J. Asl
Mohammad H. Rafiei
Manar Alohaly
Daniel Takabi
AAML
SILM
31
3
0
18 Mar 2024
RobustSentEmbed: Robust Sentence Embeddings Using Adversarial Self-Supervised Contrastive Learning
J. Asl
Prajwal Panzade
Eduardo Blanco
Daniel Takabi
Zhipeng Cai
SSL
29
2
0
17 Mar 2024
Extreme Miscalibration and the Illusion of Adversarial Robustness
Vyas Raina
Samson Tan
V. Cevher
Aditya Rawal
Sheng Zha
George Karypis
AAML
41
2
0
27 Feb 2024
Unveiling Vulnerability of Self-Attention
Khai Jiet Liong
Hongqiu Wu
Haizhen Zhao
44
0
0
26 Feb 2024
RoCoIns: Enhancing Robustness of Large Language Models through Code-Style Instructions
Yuan Zhang
Xiao Wang
Zhiheng Xi
Han Xia
Tao Gui
Qi Zhang
Xuanjing Huang
51
3
0
26 Feb 2024
RITFIS: Robust input testing framework for LLMs-based intelligent software
Ming-Ming Xiao
Yan Xiao
Hai Dong
Shunhui Ji
Pengcheng Zhang
AAML
52
5
0
21 Feb 2024
Query-Based Adversarial Prompt Generation
Jonathan Hayase
Ema Borevkovic
Nicholas Carlini
Florian Tramèr
Milad Nasr
AAML
SILM
45
26
0
19 Feb 2024
Groot: Adversarial Testing for Generative Text-to-Image Models with Tree-based Semantic Transformation
Yi Liu
Guowei Yang
Gelei Deng
Feiyue Chen
Yuqi Chen
Ling Shi
Tianwei Zhang
Yang Liu
VLM
30
9
0
19 Feb 2024
Stumbling Blocks: Stress Testing the Robustness of Machine-Generated Text Detectors Under Attacks
Yichen Wang
Shangbin Feng
Abe Bohan Hou
Xiao Pu
Chao Shen
Xiaoming Liu
Yulia Tsvetkov
Tianxing He
DeLMO
48
17
0
18 Feb 2024
A Curious Case of Searching for the Correlation between Training Data and Adversarial Robustness of Transformer Textual Models
Cuong Dang
Dung D. Le
Thai Le
AAML
37
2
0
18 Feb 2024
Pixel Sentence Representation Learning
Chenghao Xiao
Zhuoxu Huang
Danlu Chen
G. Hudson
Yizhi Li
Haoran Duan
Chenghua Lin
Jie Fu
Jungong Han
Noura Al Moubayed
SSL
17
2
0
13 Feb 2024
PoisonedRAG: Knowledge Poisoning Attacks to Retrieval-Augmented Generation of Large Language Models
Wei Zou
Runpeng Geng
Binghui Wang
Jinyuan Jia
SILM
39
19
1
12 Feb 2024
HarmBench: A Standardized Evaluation Framework for Automated Red Teaming and Robust Refusal
Mantas Mazeika
Long Phan
Xuwang Yin
Andy Zou
Zifan Wang
...
Nathaniel Li
Steven Basart
Bo Li
David A. Forsyth
Dan Hendrycks
AAML
32
327
0
06 Feb 2024
Arabic Synonym BERT-based Adversarial Examples for Text Classification
Norah M. Alshahrani
Saied Alshahrani
Esma Wali
Jeanna Neefe Matthews
AAML
22
5
0
05 Feb 2024
Data Poisoning for In-context Learning
Pengfei He
Han Xu
Yue Xing
Hui Liu
Makoto Yamada
Jiliang Tang
AAML
SILM
30
10
0
03 Feb 2024
Building Guardrails for Large Language Models
Yizhen Dong
Ronghui Mu
Gao Jin
Yi Qi
Jinwei Hu
Xingyu Zhao
Jie Meng
Wenjie Ruan
Xiaowei Huang
OffRL
63
27
0
02 Feb 2024
HQA-Attack: Toward High Quality Black-Box Hard-Label Adversarial Attack on Text
Han Liu
Zhi Xu
Xiaotong Zhang
Feng Zhang
Fenglong Ma
Hongyang Chen
Hong Yu
Xianchao Zhang
AAML
27
7
0
02 Feb 2024
Trustworthy Distributed AI Systems: Robustness, Privacy, and Governance
Wenqi Wei
Ling Liu
31
16
0
02 Feb 2024
Black-Box Access is Insufficient for Rigorous AI Audits
Stephen Casper
Carson Ezell
Charlotte Siegmann
Noam Kolt
Taylor Lynn Curtis
...
Michael Gerovitch
David Bau
Max Tegmark
David M. Krueger
Dylan Hadfield-Menell
AAML
36
78
0
25 Jan 2024
Fast Adversarial Training against Textual Adversarial Attacks
Yichen Yang
Xin Liu
Kun He
AAML
16
4
0
23 Jan 2024
Adapters Mixup: Mixing Parameter-Efficient Adapters to Enhance the Adversarial Robustness of Fine-tuned Pre-trained Text Classifiers
Tuc Nguyen
Thai Le
AAML
SILM
MoE
16
2
0
18 Jan 2024
AdvSQLi: Generating Adversarial SQL Injections against Real-world WAF-as-a-service
Zhenqing Qu
Xiang Ling
Ting Wang
Xiang Chen
Shouling Ji
Chunming Wu
AAML
24
10
0
05 Jan 2024
Silent Guardian: Protecting Text from Malicious Exploitation by Large Language Models
Jiawei Zhao
Kejiang Chen
Xianjian Yuan
Yuang Qi
Weiming Zhang
Neng H. Yu
67
8
0
15 Dec 2023
Tree of Attacks: Jailbreaking Black-Box LLMs Automatically
Anay Mehrotra
Manolis Zampetakis
Paul Kassianik
Blaine Nelson
Hyrum Anderson
Yaron Singer
Amin Karbasi
44
206
0
04 Dec 2023
SenTest: Evaluating Robustness of Sentence Encoders
Tanmay Chavan
Shantanu Patankar
Aditya Kane
Omkar Gokhale
Geetanjali Kale
Raviraj Joshi
24
0
0
29 Nov 2023
Improving the Robustness of Transformer-based Large Language Models with Dynamic Attention
Lujia Shen
Yuwen Pu
Shouling Ji
Changjiang Li
Xuhong Zhang
Chunpeng Ge
Ting Wang
AAML
29
3
0
29 Nov 2023
Generating Valid and Natural Adversarial Examples with Large Language Models
Zimu Wang
Wei Wang
Qi Chen
Qiufeng Wang
Anh Nguyen
AAML
21
4
0
20 Nov 2023
Previous
1
2
3
4
5
6
7
8
Next