Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2106.09898
Cited By
Bad Characters: Imperceptible NLP Attacks
18 June 2021
Nicholas Boucher
Ilia Shumailov
Ross J. Anderson
Nicolas Papernot
AAML
SILM
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Bad Characters: Imperceptible NLP Attacks"
50 / 68 papers shown
Title
Sponge Attacks on Sensing AI: Energy-Latency Vulnerabilities and Defense via Model Pruning
Syed Mhamudul Hasan
Hussein Zangoti
Iraklis Anagnostopoulos
Abdur R. Shahid
AAML
29
0
0
09 May 2025
A Comprehensive Analysis of Adversarial Attacks against Spam Filters
Esra Hotoğlu
Sevil Sen
Burcu Can
AAML
29
0
0
04 May 2025
Exploring Gradient-Guided Masked Language Model to Detect Textual Adversarial Attacks
Xiaomei Zhang
Zhaoxi Zhang
Yanjun Zhang
Xufei Zheng
L. Zhang
Shengshan Hu
Shirui Pan
AAML
27
0
0
08 Apr 2025
A Grey-box Text Attack Framework using Explainable AI
Esther Chiramal
Kelvin Soh Boon Kai
AAML
SILM
57
0
0
11 Mar 2025
BiMarker: Enhancing Text Watermark Detection for Large Language Models with Bipolar Watermarks
Zhuang Li
50
1
0
21 Jan 2025
The Best Defense is a Good Offense: Countering LLM-Powered Cyberattacks
Daniel Ayzenshteyn
Roy Weiss
Yisroel Mirsky
AAML
31
0
0
20 Oct 2024
Golyadkin's Torment: Doppelgängers and Adversarial Vulnerability
George I. Kamberov
AAML
29
0
0
17 Oct 2024
Probing the Robustness of Vision-Language Pretrained Models: A Multimodal Adversarial Attack Approach
Jiwei Guan
Tianyu Ding
Longbing Cao
Lei Pan
Chen Wang
Xi Zheng
AAML
33
1
0
24 Aug 2024
Breaking Agents: Compromising Autonomous LLM Agents Through Malfunction Amplification
Boyang Zhang
Yicong Tan
Yun Shen
Ahmed Salem
Michael Backes
Savvas Zannettou
Yang Zhang
LLMAG
AAML
44
14
0
30 Jul 2024
Watermark Smoothing Attacks against Language Models
Hongyan Chang
Hamed Hassani
Reza Shokri
WaLM
65
2
0
19 Jul 2024
SpamDam: Towards Privacy-Preserving and Adversary-Resistant SMS Spam Detection
Yekai Li
Rufan Zhang
Wenxin Rong
Xianghang Mi
39
2
0
15 Apr 2024
The Impact of Uniform Inputs on Activation Sparsity and Energy-Latency Attacks in Computer Vision
Andreas Müller
Erwin Quiring
AAML
34
1
0
27 Mar 2024
A Modified Word Saliency-Based Adversarial Attack on Text Classification Models
Hetvi Waghela
Sneha Rakshit
Jaydip Sen
AAML
23
7
0
17 Mar 2024
Architectural Neural Backdoors from First Principles
Harry Langford
Ilia Shumailov
Yiren Zhao
Robert D. Mullins
Nicolas Papernot
AAML
38
2
0
10 Feb 2024
Comprehensive Assessment of Jailbreak Attacks Against LLMs
Junjie Chu
Yugeng Liu
Ziqing Yang
Xinyue Shen
Michael Backes
Yang Zhang
AAML
37
66
0
08 Feb 2024
DeSparsify: Adversarial Attack Against Token Sparsification Mechanisms in Vision Transformers
Oryan Yehezkel
Alon Zolfi
Amit Baras
Yuval Elovici
A. Shabtai
AAML
32
0
0
04 Feb 2024
Silent Guardian: Protecting Text from Malicious Exploitation by Large Language Models
Jiawei Zhao
Kejiang Chen
Xianjian Yuan
Yuang Qi
Weiming Zhang
Neng H. Yu
59
8
0
15 Dec 2023
Beyond Boundaries: A Comprehensive Survey of Transferable Attacks on AI Systems
Guangjing Wang
Ce Zhou
Yuanda Wang
Bocheng Chen
Hanqing Guo
Qiben Yan
AAML
SILM
63
3
0
20 Nov 2023
BERT Lost Patience Won't Be Robust to Adversarial Slowdown
Zachary Coalson
Gabriel Ritter
Rakesh Bobba
Sanghyun Hong
AAML
24
1
0
29 Oct 2023
Invisible Threats: Backdoor Attack in OCR Systems
Mauro Conti
Nicola Farronato
Stefanos Koffas
Luca Pajola
S. Picek
32
1
0
12 Oct 2023
Jailbreak and Guard Aligned Language Models with Only Few In-Context Demonstrations
Zeming Wei
Yifei Wang
Ang Li
Yichuan Mo
Yisen Wang
45
236
0
10 Oct 2023
Enhancing Robust Representation in Adversarial Training: Alignment and Exclusion Criteria
Nuoyan Zhou
Nannan Wang
Decheng Liu
Dawei Zhou
Xinbo Gao
AAML
33
2
0
05 Oct 2023
WASA: WAtermark-based Source Attribution for Large Language Model-Generated Data
Jingtan Wang
Xinyang Lu
Zitong Zhao
Zhongxiang Dai
Chuan-Sheng Foo
See-Kiong Ng
K. H. Low
WaLM
57
14
0
01 Oct 2023
Unbiased Watermark for Large Language Models
Zhengmian Hu
Lichang Chen
Xidong Wu
Yihan Wu
Hongyang R. Zhang
Heng-Chiao Huang
WaLM
38
45
0
22 Sep 2023
"Do Anything Now": Characterizing and Evaluating In-The-Wild Jailbreak Prompts on Large Language Models
Xinyue Shen
Z. Chen
Michael Backes
Yun Shen
Yang Zhang
SILM
40
245
0
07 Aug 2023
LLM Censorship: A Machine Learning Challenge or a Computer Security Problem?
David Glukhov
Ilia Shumailov
Y. Gal
Nicolas Papernot
V. Papyan
AAML
ELM
26
57
0
20 Jul 2023
Efficient Parallel Output-Sensitive Edit Distance
Xiangyun Ding
Xiaojun Dong
Yan Gu
Youzhe Liu
Yihan Sun
20
4
0
30 Jun 2023
When Vision Fails: Text Attacks Against ViT and OCR
Nicholas Boucher
Jenny Blessing
Ilia Shumailov
Ross J. Anderson
Nicolas Papernot
AAML
34
4
0
12 Jun 2023
Enhancing Robustness of AI Offensive Code Generators via Data Augmentation
Cristina Improta
Pietro Liguori
R. Natella
B. Cukic
Domenico Cotroneo
AAML
30
2
0
08 Jun 2023
Avoid Adversarial Adaption in Federated Learning by Multi-Metric Investigations
T. Krauß
Alexandra Dmitrienko
AAML
27
4
0
06 Jun 2023
Two-in-One: A Model Hijacking Attack Against Text Generation Models
Waiman Si
Michael Backes
Yang Zhang
A. Salem
SILM
24
22
0
12 May 2023
Boosting Big Brother: Attacking Search Engines with Encodings
Nicholas Boucher
Luca Pajola
Ilia Shumailov
Ross J. Anderson
Mauro Conti
SILM
29
10
0
27 Apr 2023
RoCOCO: Robustness Benchmark of MS-COCO to Stress-test Image-Text Matching Models
Seulki Park
Daeho Um
Hajung Yoon
Sanghyuk Chun
Sangdoo Yun
Jin Young Choi
38
2
0
21 Apr 2023
RNN-Guard: Certified Robustness Against Multi-frame Attacks for Recurrent Neural Networks
Yunruo Zhang
Tianyu Du
S. Ji
Peng Tang
Shanqing Guo
AAML
31
2
0
17 Apr 2023
No more Reviewer #2: Subverting Automatic Paper-Reviewer Assignment using Adversarial Learning
Thorsten Eisenhofer
Erwin Quiring
Jonas Moller
Doreen Riepel
Thorsten Holz
Konrad Rieck
AAML
26
6
0
25 Mar 2023
Model-tuning Via Prompts Makes NLP Models Adversarially Robust
Mrigank Raman
Pratyush Maini
J. Zico Kolter
Zachary Chase Lipton
Danish Pruthi
AAML
35
17
0
13 Mar 2023
Learning the Legibility of Visual Text Perturbations
D. Seth
Rickard Stureborg
Danish Pruthi
Bhuwan Dhingra
AAML
48
4
0
09 Mar 2023
Not what you've signed up for: Compromising Real-World LLM-Integrated Applications with Indirect Prompt Injection
Kai Greshake
Sahar Abdelnabi
Shailesh Mishra
C. Endres
Thorsten Holz
Mario Fritz
SILM
49
436
0
23 Feb 2023
MTTM: Metamorphic Testing for Textual Content Moderation Software
Wenxuan Wang
Jen-tse Huang
Weibin Wu
Jianping Zhang
Yizhan Huang
Shuqing Li
Pinjia He
Michael Lyu
50
29
0
11 Feb 2023
Training-free Lexical Backdoor Attacks on Language Models
Yujin Huang
Terry Yue Zhuo
Qiongkai Xu
Han Hu
Xingliang Yuan
Chunyang Chen
SILM
28
42
0
08 Feb 2023
On Robustness of Prompt-based Semantic Parsing with Large Pre-trained Language Model: An Empirical Study on Codex
Terry Yue Zhuo
Zhuang Li
Yujin Huang
Fatemeh Shiri
Weiqing Wang
Gholamreza Haffari
Yuan-Fang Li
AAML
26
53
0
30 Jan 2023
Demystifying Privacy Policy of Third-Party Libraries in Mobile Apps
Kaifa Zhao
Xian Zhan
Le Yu
Shiyao Zhou
Hao Zhou
Xiapu Luo
Haoyu Wang
Yepang Liu
37
14
0
29 Jan 2023
A Watermark for Large Language Models
John Kirchenbauer
Jonas Geiping
Yuxin Wen
Jonathan Katz
Ian Miers
Tom Goldstein
VLM
WaLM
20
460
0
24 Jan 2023
DDoD: Dual Denial of Decision Attacks on Human-AI Teams
Benjamin Tag
N. V. Berkel
Sunny Verma
Benjamin Zi Hao Zhao
S. Berkovsky
Dali Kaafar
V. Kostakos
O. Ohrimenko
AAML
25
5
0
07 Dec 2022
On the Security Vulnerabilities of Text-to-SQL Models
Xutan Peng
Yipeng Zhang
Jingfeng Yang
Mark Stevenson
SILM
23
10
0
28 Nov 2022
Secure and Trustworthy Artificial Intelligence-Extended Reality (AI-XR) for Metaverses
Adnan Qayyum
M. A. Butt
Hassan Ali
Muhammad Usman
O. Halabi
Ala I. Al-Fuqaha
Q. Abbasi
Muhammad Ali Imran
Junaid Qadir
30
32
0
24 Oct 2022
ImpNet: Imperceptible and blackbox-undetectable backdoors in compiled neural networks
Eleanor Clifford
Ilia Shumailov
Yiren Zhao
Ross J. Anderson
Robert D. Mullins
23
12
0
30 Sep 2022
Talking Trojan: Analyzing an Industry-Wide Disclosure
Nicholas Boucher
Ross J. Anderson
24
3
0
22 Sep 2022
Exploiting Cultural Biases via Homoglyphs in Text-to-Image Synthesis
Lukas Struppek
Dominik Hintersdorf
Felix Friedrich
Manuel Brack
P. Schramowski
Kristian Kersting
73
26
0
19 Sep 2022
Fact-Saboteurs: A Taxonomy of Evidence Manipulation Attacks against Fact-Verification Systems
Sahar Abdelnabi
Mario Fritz
AAML
197
5
0
07 Sep 2022
1
2
Next