ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1802.00420
  4. Cited By
Obfuscated Gradients Give a False Sense of Security: Circumventing
  Defenses to Adversarial Examples
v1v2v3v4 (latest)

Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples

1 February 2018
Anish Athalye
Nicholas Carlini
D. Wagner
    AAML
ArXiv (abs)PDFHTML

Papers citing "Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples"

50 / 1,929 papers shown
Title
RL-Obfuscation: Can Language Models Learn to Evade Latent-Space Monitors?
RL-Obfuscation: Can Language Models Learn to Evade Latent-Space Monitors?
Rohan Gupta
Erik Jenner
27
0
0
17 Jun 2025
Busting the Paper Ballot: Voting Meets Adversarial Machine Learning
Busting the Paper Ballot: Voting Meets Adversarial Machine Learning
Kaleel Mahmood
Caleb Manicke
Ethan Rathbun
Aayushi Verma
Sohaib Ahmad
Nicholas Stamatakis
L. Michel
Benjamin Fuller
AAML
45
0
0
17 Jun 2025
Position: Certified Robustness Does Not (Yet) Imply Model Security
Position: Certified Robustness Does Not (Yet) Imply Model Security
Andrew C. Cullen
Paul Montague
S. Erfani
Benjamin I. P. Rubinstein
26
0
0
16 Jun 2025
Existence of Adversarial Examples for Random Convolutional Networks via Isoperimetric Inequalities on $\mathbb{so}(d)$
Existence of Adversarial Examples for Random Convolutional Networks via Isoperimetric Inequalities on so(d)\mathbb{so}(d)so(d)
Amit Daniely
28
0
0
14 Jun 2025
Attention-based Adversarial Robust Distillation in Radio Signal Classifications for Low-Power IoT Devices
Attention-based Adversarial Robust Distillation in Radio Signal Classifications for Low-Power IoT Devices
Lu Zhang
S. Lambotharan
G. Zheng
G. Liao
Basil AsSadhan
Fabio Roli
AAML
21
7
0
13 Jun 2025
Lattice Climber Attack: Adversarial attacks for randomized mixtures of classifiers
Lattice Climber Attack: Adversarial attacks for randomized mixtures of classifiers
Lucas Gnecco-Heredia
Benjamin Négrevergne
Y. Chevaleyre
AAML
109
0
0
12 Jun 2025
A Crack in the Bark: Leveraging Public Knowledge to Remove Tree-Ring Watermarks
A Crack in the Bark: Leveraging Public Knowledge to Remove Tree-Ring Watermarks
Junhua Lin
Marc Juarez
116
0
0
12 Jun 2025
SHIELD: Secure Hypernetworks for Incremental Expansion Learning Defense
Patryk Krukowski
Łukasz Gorczyca
Piotr Helm
Kamil Ksiazek
Przemysław Spurek
AAMLCLL
30
0
0
09 Jun 2025
PASS: Private Attributes Protection with Stochastic Data Substitution
PASS: Private Attributes Protection with Stochastic Data Substitution
Yizhuo Chen
Chun-Fu
Chen
Hsiang Hsu
Shaohan Hu
Tarek Abdelzaher
23
0
0
08 Jun 2025
Sample-Specific Noise Injection For Diffusion-Based Adversarial Purification
Sample-Specific Noise Injection For Diffusion-Based Adversarial Purification
Yuhao Sun
Jiacheng Zhang
Zesheng Ye
Chaowei Xiao
Feng Liu
DiffM
50
0
0
06 Jun 2025
SDN-Based False Data Detection With Its Mitigation and Machine Learning Robustness for In-Vehicle Networks
SDN-Based False Data Detection With Its Mitigation and Machine Learning Robustness for In-Vehicle Networks
Long Dang
T. Hapuarachchi
Kaiqi Xiong
Yi Li
AAML
18
0
0
06 Jun 2025
Identifying and Understanding Cross-Class Features in Adversarial Training
Zeming Wei
Yiwen Guo
Yisen Wang
AAML
102
0
0
05 Jun 2025
Efficient Robust Conformal Prediction via Lipschitz-Bounded Networks
Efficient Robust Conformal Prediction via Lipschitz-Bounded Networks
Thomas Massena
Léo Andéol
Thibaut Boissin
Franck Mamalet
Corentin Friedrich
M. Serrurier
Sébastien Gerchinovitz
AAML
45
2
0
05 Jun 2025
Fighting Fire with Fire (F3): A Training-free and Efficient Visual Adversarial Example Purification Method in LVLMs
Fighting Fire with Fire (F3): A Training-free and Efficient Visual Adversarial Example Purification Method in LVLMs
Yudong Zhang
Ruobing Xie
Yiqing Huang
Jiansheng Chen
Xingwu Sun
Zhanhui Kang
Di Wang
Yu Wang
AAML
49
0
0
01 Jun 2025
Differential Privacy for Deep Learning in Medicine
Differential Privacy for Deep Learning in Medicine
Marziyeh Mohammadi
Mohsen Vejdanihemmat
Mahshad Lotfinia
M. Rusu
Daniel Truhn
Andreas K. Maier
Soroosh Tayebi Arasteh
42
0
0
31 May 2025
How Do Diffusion Models Improve Adversarial Robustness?
How Do Diffusion Models Improve Adversarial Robustness?
Liu Yuezhang
Xue-Xin Wei
296
0
0
28 May 2025
LORE: Lagrangian-Optimized Robust Embeddings for Visual Encoders
LORE: Lagrangian-Optimized Robust Embeddings for Visual Encoders
Borna Khodabandeh
Amirabbas Afzali
Amirhossein Afsharrad
Seyed Shahabeddin Mousavi
Sanjay Lall
Sajjad Amini
Seyed-Mohsen Moosavi-Dezfooli
AAML
38
0
0
24 May 2025
Beyond Discreteness: Finite-Sample Analysis of Straight-Through Estimator for Quantization
Halyun Jeong
Jack Xin
Penghang Yin
MQ
37
0
0
23 May 2025
My Face Is Mine, Not Yours: Facial Protection Against Diffusion Model Face Swapping
My Face Is Mine, Not Yours: Facial Protection Against Diffusion Model Face Swapping
Hon Ming Yam
Zhongliang Guo
Chun Pong Lau
DiffMAAML
60
0
0
21 May 2025
Adversarially Pretrained Transformers may be Universally Robust In-Context Learners
Adversarially Pretrained Transformers may be Universally Robust In-Context Learners
Soichiro Kumano
Hiroshi Kera
Toshihiko Yamasaki
AAML
127
0
0
20 May 2025
FlowPure: Continuous Normalizing Flows for Adversarial Purification
FlowPure: Continuous Normalizing Flows for Adversarial Purification
Elias Collaert
Abel Rodríguez
Sander Joos
Lieven Desmet
Vera Rimmer
AAML
67
0
0
19 May 2025
Adversarially Robust Spiking Neural Networks with Sparse Connectivity
Adversarially Robust Spiking Neural Networks with Sparse Connectivity
Mathias Schmolli
Maximilian Baronig
Robert Legenstein
Ozan Özdenizci
AAML
45
0
0
16 May 2025
TAROT: Towards Essentially Domain-Invariant Robustness with Theoretical Justification
TAROT: Towards Essentially Domain-Invariant Robustness with Theoretical Justification
Dongyoon Yang
Jihu Lee
Yongdai Kim
99
0
0
10 May 2025
Diffusion-based Adversarial Purification from the Perspective of the Frequency Domain
Diffusion-based Adversarial Purification from the Perspective of the Frequency Domain
Gaozheng Pei
Ke Ma
Yingfei Sun
Qianqian Xu
Qingming Huang
DiffM
84
0
0
02 May 2025
Edge-Based Learning for Improved Classification Under Adversarial Noise
Edge-Based Learning for Improved Classification Under Adversarial Noise
Manish Kansana
Keyan Alexander Rahimi
Elias Hossain
Iman Dehzangi
Noorbakhsh Amiri Golilarz
AAML
63
0
0
25 Apr 2025
Towards Robust LLMs: an Adversarial Robustness Measurement Framework
Towards Robust LLMs: an Adversarial Robustness Measurement Framework
Natan Levy
Adiel Ashrov
Guy Katz
AAML
83
0
0
24 Apr 2025
Adversarial Examples in Environment Perception for Automated Driving (Review)
Adversarial Examples in Environment Perception for Automated Driving (Review)
Jun Yan
Huilin Yin
AAML
93
0
0
11 Apr 2025
Mind the Trojan Horse: Image Prompt Adapter Enabling Scalable and Deceptive Jailbreaking
Mind the Trojan Horse: Image Prompt Adapter Enabling Scalable and Deceptive Jailbreaking
Junxi Chen
Junhao Dong
Xiaohua Xie
89
0
0
08 Apr 2025
A Study on Adversarial Robustness of Discriminative Prototypical Learning
A Study on Adversarial Robustness of Discriminative Prototypical Learning
Ramin Zarei-Sabzevar
Hamed Mohammadzadeh
Tahmineh Tavakoli
Ahad Harati
AAML
89
0
0
03 Apr 2025
Revisiting the Relationship between Adversarial and Clean Training: Why Clean Training Can Make Adversarial Training Better
Revisiting the Relationship between Adversarial and Clean Training: Why Clean Training Can Make Adversarial Training Better
MingWei Zhou
Xiaobing Pei
AAML
449
0
0
30 Mar 2025
Stop Walking in Circles! Bailing Out Early in Projected Gradient Descent
Stop Walking in Circles! Bailing Out Early in Projected Gradient Descent
Philip Doldo
Derek Everett
Amol Khanna
A. Nguyen
Edward Raff
AAML
85
0
0
25 Mar 2025
Principal Eigenvalue Regularization for Improved Worst-Class Certified Robustness of Smoothed Classifiers
Principal Eigenvalue Regularization for Improved Worst-Class Certified Robustness of Smoothed Classifiers
Gaojie Jin
Tianjin Huang
Ronghui Mu
Xiaowei Huang
AAML
77
0
0
21 Mar 2025
Narrowing Class-Wise Robustness Gaps in Adversarial Training
Narrowing Class-Wise Robustness Gaps in Adversarial Training
Fatemeh Amerehi
Patrick Healy
101
0
0
20 Mar 2025
Robust Dataset Distillation by Matching Adversarial Trajectories
Robust Dataset Distillation by Matching Adversarial Trajectories
Wei Lai
Tianyu Ding
ren dongdong
Lei Wang
Jing Huo
Yang Gao
Wenbin Li
AAMLDD
102
0
0
15 Mar 2025
Are Deep Speech Denoising Models Robust to Adversarial Noise?
Will Schwarzer
Philip S. Thomas
Andrea Fanelli
Xiaoyu Liu
75
0
0
14 Mar 2025
Provenance Detection for AI-Generated Images: Combining Perceptual Hashing, Homomorphic Encryption, and AI Detection Models
Shree Singhi
Aayan Yadav
Aayush Gupta
Shariar Ebrahimi
Parisa Hassanizadeh
88
1
0
14 Mar 2025
Attacking Multimodal OS Agents with Malicious Image Patches
Lukas Aichberger
Alasdair Paren
Y. Gal
Philip Torr
Adel Bibi
AAML
121
5
0
13 Mar 2025
Probabilistic Segmentation for Robust Field of View Estimation
R. S. Hallyburton
David Hunt
Yiwei He
Judy He
Miroslav Pajic
107
1
0
10 Mar 2025
Breaking the Limits of Quantization-Aware Defenses: QADT-R for Robustness Against Patch-Based Adversarial Attacks in QNNs
Amira Guesmi
B. Ouni
Muhammad Shafique
MQAAML
132
0
0
10 Mar 2025
Long-tailed Adversarial Training with Self-Distillation
Seungju Cho
Hongsin Lee
Changick Kim
AAMLTTA
498
0
0
09 Mar 2025
MMARD: Improving the Min-Max Optimization Process in Adversarial Robustness Distillation
Yuzheng Wang
Zhaoyu Chen
Dingkang Yang
Yuanhang Wang
Lizhe Qi
AAML
147
0
0
09 Mar 2025
CLIP is Strong Enough to Fight Back: Test-time Counterattacks towards Zero-shot Adversarial Robustness of CLIP
Songlong Xing
Zhengyu Zhao
N. Sebe
AAML
155
2
0
05 Mar 2025
LLM-Safety Evaluations Lack Robustness
Tim Beyer
Sophie Xhonneux
Simon Geisler
Gauthier Gidel
Leo Schwinn
Stephan Günnemann
ALMELM
485
2
0
04 Mar 2025
One Stone, Two Birds: Enhancing Adversarial Defense Through the Lens of Distributional Discrepancy
One Stone, Two Birds: Enhancing Adversarial Defense Through the Lens of Distributional Discrepancy
Jiacheng Zhang
Benjamin I. P. Rubinstein
Jing Zhang
Feng Liu
131
0
0
04 Mar 2025
AutoAdvExBench: Benchmarking autonomous exploitation of adversarial example defenses
Nicholas Carlini
Javier Rando
Edoardo Debenedetti
Milad Nasr
F. Tramèr
AAMLELM
92
3
0
03 Mar 2025
Fast Adversarial Training against Sparse Attacks Requires Loss Smoothing
Fast Adversarial Training against Sparse Attacks Requires Loss Smoothing
Xuyang Zhong
Yixiao Huang
Chen Liu
AAML
125
0
0
28 Feb 2025
HALO: Robust Out-of-Distribution Detection via Joint Optimisation
HALO: Robust Out-of-Distribution Detection via Joint Optimisation
Hugo Lyons Keenan
S. Erfani
Christopher Leckie
OODD
276
0
0
27 Feb 2025
CLIPure: Purification in Latent Space via CLIP for Adversarially Robust Zero-Shot Classification
CLIPure: Purification in Latent Space via CLIP for Adversarially Robust Zero-Shot Classification
Mingkun Zhang
Keping Bi
Wei Chen
Jiafeng Guo
Xueqi Cheng
BDLVLM
174
2
0
25 Feb 2025
REINFORCE Adversarial Attacks on Large Language Models: An Adaptive, Distributional, and Semantic Objective
Simon Geisler
Tom Wollschlager
M. H. I. Abdalla
Vincent Cohen-Addad
Johannes Gasteiger
Stephan Günnemann
AAML
128
3
0
24 Feb 2025
SEA: Shareable and Explainable Attribution for Query-based Black-box Attacks
SEA: Shareable and Explainable Attribution for Query-based Black-box Attacks
Yue Gao
Ilia Shumailov
Kassem Fawaz
AAML
222
0
0
21 Feb 2025
1234...373839
Next