Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1802.00420
Cited By
v1
v2
v3
v4 (latest)
Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples
1 February 2018
Anish Athalye
Nicholas Carlini
D. Wagner
AAML
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples"
50 / 1,929 papers shown
Title
RL-Obfuscation: Can Language Models Learn to Evade Latent-Space Monitors?
Rohan Gupta
Erik Jenner
27
0
0
17 Jun 2025
Busting the Paper Ballot: Voting Meets Adversarial Machine Learning
Kaleel Mahmood
Caleb Manicke
Ethan Rathbun
Aayushi Verma
Sohaib Ahmad
Nicholas Stamatakis
L. Michel
Benjamin Fuller
AAML
45
0
0
17 Jun 2025
Position: Certified Robustness Does Not (Yet) Imply Model Security
Andrew C. Cullen
Paul Montague
S. Erfani
Benjamin I. P. Rubinstein
26
0
0
16 Jun 2025
Existence of Adversarial Examples for Random Convolutional Networks via Isoperimetric Inequalities on
s
o
(
d
)
\mathbb{so}(d)
so
(
d
)
Amit Daniely
28
0
0
14 Jun 2025
Attention-based Adversarial Robust Distillation in Radio Signal Classifications for Low-Power IoT Devices
Lu Zhang
S. Lambotharan
G. Zheng
G. Liao
Basil AsSadhan
Fabio Roli
AAML
21
7
0
13 Jun 2025
Lattice Climber Attack: Adversarial attacks for randomized mixtures of classifiers
Lucas Gnecco-Heredia
Benjamin Négrevergne
Y. Chevaleyre
AAML
109
0
0
12 Jun 2025
A Crack in the Bark: Leveraging Public Knowledge to Remove Tree-Ring Watermarks
Junhua Lin
Marc Juarez
116
0
0
12 Jun 2025
SHIELD: Secure Hypernetworks for Incremental Expansion Learning Defense
Patryk Krukowski
Łukasz Gorczyca
Piotr Helm
Kamil Ksiazek
Przemysław Spurek
AAML
CLL
30
0
0
09 Jun 2025
PASS: Private Attributes Protection with Stochastic Data Substitution
Yizhuo Chen
Chun-Fu
Chen
Hsiang Hsu
Shaohan Hu
Tarek Abdelzaher
23
0
0
08 Jun 2025
Sample-Specific Noise Injection For Diffusion-Based Adversarial Purification
Yuhao Sun
Jiacheng Zhang
Zesheng Ye
Chaowei Xiao
Feng Liu
DiffM
50
0
0
06 Jun 2025
SDN-Based False Data Detection With Its Mitigation and Machine Learning Robustness for In-Vehicle Networks
Long Dang
T. Hapuarachchi
Kaiqi Xiong
Yi Li
AAML
18
0
0
06 Jun 2025
Identifying and Understanding Cross-Class Features in Adversarial Training
Zeming Wei
Yiwen Guo
Yisen Wang
AAML
102
0
0
05 Jun 2025
Efficient Robust Conformal Prediction via Lipschitz-Bounded Networks
Thomas Massena
Léo Andéol
Thibaut Boissin
Franck Mamalet
Corentin Friedrich
M. Serrurier
Sébastien Gerchinovitz
AAML
45
2
0
05 Jun 2025
Fighting Fire with Fire (F3): A Training-free and Efficient Visual Adversarial Example Purification Method in LVLMs
Yudong Zhang
Ruobing Xie
Yiqing Huang
Jiansheng Chen
Xingwu Sun
Zhanhui Kang
Di Wang
Yu Wang
AAML
49
0
0
01 Jun 2025
Differential Privacy for Deep Learning in Medicine
Marziyeh Mohammadi
Mohsen Vejdanihemmat
Mahshad Lotfinia
M. Rusu
Daniel Truhn
Andreas K. Maier
Soroosh Tayebi Arasteh
42
0
0
31 May 2025
How Do Diffusion Models Improve Adversarial Robustness?
Liu Yuezhang
Xue-Xin Wei
296
0
0
28 May 2025
LORE: Lagrangian-Optimized Robust Embeddings for Visual Encoders
Borna Khodabandeh
Amirabbas Afzali
Amirhossein Afsharrad
Seyed Shahabeddin Mousavi
Sanjay Lall
Sajjad Amini
Seyed-Mohsen Moosavi-Dezfooli
AAML
38
0
0
24 May 2025
Beyond Discreteness: Finite-Sample Analysis of Straight-Through Estimator for Quantization
Halyun Jeong
Jack Xin
Penghang Yin
MQ
37
0
0
23 May 2025
My Face Is Mine, Not Yours: Facial Protection Against Diffusion Model Face Swapping
Hon Ming Yam
Zhongliang Guo
Chun Pong Lau
DiffM
AAML
60
0
0
21 May 2025
Adversarially Pretrained Transformers may be Universally Robust In-Context Learners
Soichiro Kumano
Hiroshi Kera
Toshihiko Yamasaki
AAML
127
0
0
20 May 2025
FlowPure: Continuous Normalizing Flows for Adversarial Purification
Elias Collaert
Abel Rodríguez
Sander Joos
Lieven Desmet
Vera Rimmer
AAML
67
0
0
19 May 2025
Adversarially Robust Spiking Neural Networks with Sparse Connectivity
Mathias Schmolli
Maximilian Baronig
Robert Legenstein
Ozan Özdenizci
AAML
45
0
0
16 May 2025
TAROT: Towards Essentially Domain-Invariant Robustness with Theoretical Justification
Dongyoon Yang
Jihu Lee
Yongdai Kim
99
0
0
10 May 2025
Diffusion-based Adversarial Purification from the Perspective of the Frequency Domain
Gaozheng Pei
Ke Ma
Yingfei Sun
Qianqian Xu
Qingming Huang
DiffM
84
0
0
02 May 2025
Edge-Based Learning for Improved Classification Under Adversarial Noise
Manish Kansana
Keyan Alexander Rahimi
Elias Hossain
Iman Dehzangi
Noorbakhsh Amiri Golilarz
AAML
63
0
0
25 Apr 2025
Towards Robust LLMs: an Adversarial Robustness Measurement Framework
Natan Levy
Adiel Ashrov
Guy Katz
AAML
83
0
0
24 Apr 2025
Adversarial Examples in Environment Perception for Automated Driving (Review)
Jun Yan
Huilin Yin
AAML
93
0
0
11 Apr 2025
Mind the Trojan Horse: Image Prompt Adapter Enabling Scalable and Deceptive Jailbreaking
Junxi Chen
Junhao Dong
Xiaohua Xie
89
0
0
08 Apr 2025
A Study on Adversarial Robustness of Discriminative Prototypical Learning
Ramin Zarei-Sabzevar
Hamed Mohammadzadeh
Tahmineh Tavakoli
Ahad Harati
AAML
89
0
0
03 Apr 2025
Revisiting the Relationship between Adversarial and Clean Training: Why Clean Training Can Make Adversarial Training Better
MingWei Zhou
Xiaobing Pei
AAML
449
0
0
30 Mar 2025
Stop Walking in Circles! Bailing Out Early in Projected Gradient Descent
Philip Doldo
Derek Everett
Amol Khanna
A. Nguyen
Edward Raff
AAML
85
0
0
25 Mar 2025
Principal Eigenvalue Regularization for Improved Worst-Class Certified Robustness of Smoothed Classifiers
Gaojie Jin
Tianjin Huang
Ronghui Mu
Xiaowei Huang
AAML
77
0
0
21 Mar 2025
Narrowing Class-Wise Robustness Gaps in Adversarial Training
Fatemeh Amerehi
Patrick Healy
101
0
0
20 Mar 2025
Robust Dataset Distillation by Matching Adversarial Trajectories
Wei Lai
Tianyu Ding
ren dongdong
Lei Wang
Jing Huo
Yang Gao
Wenbin Li
AAML
DD
102
0
0
15 Mar 2025
Are Deep Speech Denoising Models Robust to Adversarial Noise?
Will Schwarzer
Philip S. Thomas
Andrea Fanelli
Xiaoyu Liu
75
0
0
14 Mar 2025
Provenance Detection for AI-Generated Images: Combining Perceptual Hashing, Homomorphic Encryption, and AI Detection Models
Shree Singhi
Aayan Yadav
Aayush Gupta
Shariar Ebrahimi
Parisa Hassanizadeh
88
1
0
14 Mar 2025
Attacking Multimodal OS Agents with Malicious Image Patches
Lukas Aichberger
Alasdair Paren
Y. Gal
Philip Torr
Adel Bibi
AAML
121
5
0
13 Mar 2025
Probabilistic Segmentation for Robust Field of View Estimation
R. S. Hallyburton
David Hunt
Yiwei He
Judy He
Miroslav Pajic
107
1
0
10 Mar 2025
Breaking the Limits of Quantization-Aware Defenses: QADT-R for Robustness Against Patch-Based Adversarial Attacks in QNNs
Amira Guesmi
B. Ouni
Muhammad Shafique
MQ
AAML
132
0
0
10 Mar 2025
Long-tailed Adversarial Training with Self-Distillation
Seungju Cho
Hongsin Lee
Changick Kim
AAML
TTA
498
0
0
09 Mar 2025
MMARD: Improving the Min-Max Optimization Process in Adversarial Robustness Distillation
Yuzheng Wang
Zhaoyu Chen
Dingkang Yang
Yuanhang Wang
Lizhe Qi
AAML
147
0
0
09 Mar 2025
CLIP is Strong Enough to Fight Back: Test-time Counterattacks towards Zero-shot Adversarial Robustness of CLIP
Songlong Xing
Zhengyu Zhao
N. Sebe
AAML
155
2
0
05 Mar 2025
LLM-Safety Evaluations Lack Robustness
Tim Beyer
Sophie Xhonneux
Simon Geisler
Gauthier Gidel
Leo Schwinn
Stephan Günnemann
ALM
ELM
485
2
0
04 Mar 2025
One Stone, Two Birds: Enhancing Adversarial Defense Through the Lens of Distributional Discrepancy
Jiacheng Zhang
Benjamin I. P. Rubinstein
Jing Zhang
Feng Liu
131
0
0
04 Mar 2025
AutoAdvExBench: Benchmarking autonomous exploitation of adversarial example defenses
Nicholas Carlini
Javier Rando
Edoardo Debenedetti
Milad Nasr
F. Tramèr
AAML
ELM
92
3
0
03 Mar 2025
Fast Adversarial Training against Sparse Attacks Requires Loss Smoothing
Xuyang Zhong
Yixiao Huang
Chen Liu
AAML
125
0
0
28 Feb 2025
HALO: Robust Out-of-Distribution Detection via Joint Optimisation
Hugo Lyons Keenan
S. Erfani
Christopher Leckie
OODD
276
0
0
27 Feb 2025
CLIPure: Purification in Latent Space via CLIP for Adversarially Robust Zero-Shot Classification
Mingkun Zhang
Keping Bi
Wei Chen
Jiafeng Guo
Xueqi Cheng
BDL
VLM
174
2
0
25 Feb 2025
REINFORCE Adversarial Attacks on Large Language Models: An Adaptive, Distributional, and Semantic Objective
Simon Geisler
Tom Wollschlager
M. H. I. Abdalla
Vincent Cohen-Addad
Johannes Gasteiger
Stephan Günnemann
AAML
128
3
0
24 Feb 2025
SEA: Shareable and Explainable Attribution for Query-based Black-box Attacks
Yue Gao
Ilia Shumailov
Kassem Fawaz
AAML
222
0
0
21 Feb 2025
1
2
3
4
...
37
38
39
Next