Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1608.04644
Cited By
v1
v2 (latest)
Towards Evaluating the Robustness of Neural Networks
16 August 2016
Nicholas Carlini
D. Wagner
OOD
AAML
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Towards Evaluating the Robustness of Neural Networks"
50 / 4,015 papers shown
Title
Towards Imperceptible Adversarial Attacks for Time Series Classification with Local Perturbations and Frequency Analysis
Wenwei Gu
Renyi Zhong
Jianping Zhang
Michael R. Lyu
AAML
91
0
0
25 Mar 2025
Bitstream Collisions in Neural Image Compression via Adversarial Perturbations
Jordan Madden
Lhamo Dorje
Xiaohua Li
AAML
79
0
0
25 Mar 2025
Input-Triggered Hardware Trojan Attack on Spiking Neural Networks
Spyridon Raptis
Paul Kling
Ioannis Kaskampas
Ihsen Alouani
Haralampos-G. Stratigopoulos
AAML
87
1
0
20 Mar 2025
Rethinking Robustness in Machine Learning: A Posterior Agreement Approach
João B. S. Carvalho
Alessandro Torcinovich
Victor Jimenez Rodriguez
Antonio Emanuele Cinà
Carlos Cotrini
Lea Schönherr
J. M. Buhmann
OOD
111
0
0
20 Mar 2025
On the Robustness Tradeoff in Fine-Tuning
Kunyang Li
Jean-Charles Noirot Ferrand
Ryan Sheatsley
Blaine Hoak
Yohan Beugin
Eric Pauley
Patrick McDaniel
93
0
0
19 Mar 2025
AIGVE-Tool: AI-Generated Video Evaluation Toolkit with Multifaceted Benchmark
Xinhao Xiang
Xiao Liu
Zizhong Li
Zhuosheng Liu
Jiawei Zhang
93
0
0
18 Mar 2025
TarPro: Targeted Protection against Malicious Image Editing
Kaixin Shen
Ruijie Quan
Jiaxu Miao
Jun Xiao
Yi Yang
113
1
0
18 Mar 2025
Securing Virtual Reality Experiences: Unveiling and Tackling Cybersickness Attacks with Explainable AI
Ripan Kumar Kundu
Matthew Denton
Genova Mongalo
Prasad Calyam
K. A. Hoque
AAML
84
0
0
17 Mar 2025
GSBA
K
^K
K
:
t
o
p
top
t
o
p
-
K
K
K
Geometric Score-based Black-box Attack
Md. Farhamdur Reza
Richeng Jin
Tianfu Wu
H. Dai
AAML
114
0
0
17 Mar 2025
Weakly Supervised Contrastive Adversarial Training for Learning Robust Features from Semi-supervised Data
Lilin Zhang
Chengpei Wu
Ning Yang
105
0
0
14 Mar 2025
A Frustratingly Simple Yet Highly Effective Attack Baseline: Over 90% Success Rate Against the Strong Black-box Models of GPT-4.5/4o/o1
Zhaoyi Li
Xiaohan Zhao
Dong-Dong Wu
Jiacheng Cui
Zhiqiang Shen
AAML
VLM
144
3
0
13 Mar 2025
Robustness Tokens: Towards Adversarial Robustness of Transformers
Brian Pulfer
Yury Belousov
S. Voloshynovskiy
AAML
85
0
0
13 Mar 2025
AdvAD: Exploring Non-Parametric Diffusion for Imperceptible Adversarial Attacks
Jin Li
Ziqiang He
Anwei Luo
Jian-Fang Hu
Zhong Wang
Xiangui Kang
DiffM
125
0
0
12 Mar 2025
Enhancing Adversarial Example Detection Through Model Explanation
Qian Ma
Ziping Ye
AAML
100
0
0
12 Mar 2025
FairDeFace: Evaluating the Fairness and Adversarial Robustness of Face Obfuscation Methods
Seyyed Mohammad Sadegh Moosavi Khorzooghi
Poojitha Thota
Mohit Singhal
Abolfazl Asudeh
Gautam Das
Shirin Nilizadeh
AAML
73
0
0
11 Mar 2025
Trustworthy Machine Learning via Memorization and the Granular Long-Tail: A Survey on Interactions, Tradeoffs, and Beyond
Qiongxiu Li
Xiaoyu Luo
Yiyi Chen
Johannes Bjerva
242
2
0
10 Mar 2025
Long-tailed Adversarial Training with Self-Distillation
Seungju Cho
Hongsin Lee
Changick Kim
AAML
TTA
500
0
0
09 Mar 2025
MMARD: Improving the Min-Max Optimization Process in Adversarial Robustness Distillation
Yuzheng Wang
Zhaoyu Chen
Jinjie Wei
Yuanhang Wang
Lizhe Qi
AAML
149
0
0
09 Mar 2025
Life-Cycle Routing Vulnerabilities of LLM Router
Qiqi Lin
Xiaoyang Ji
Shengfang Zhai
Qingni Shen
Zhi-Li Zhang
Yuejian Fang
Yansong Gao
AAML
90
1
0
09 Mar 2025
Energy-Latency Attacks: A New Adversarial Threat to Deep Learning
H. B. Meftah
W. Hamidouche
Sid Ahmed Fezza
Olivier Déforges
AAML
72
0
0
06 Mar 2025
An Information-theoretic Multi-task Representation Learning Framework for Natural Language Understanding
Dou Hu
Lingwei Wei
Wei Zhou
Songlin Hu
125
0
0
06 Mar 2025
Task-Agnostic Attacks Against Vision Foundation Models
Brian Pulfer
Yury Belousov
Vitaliy Kinakh
Teddy Furon
S. Voloshynovskiy
AAML
111
0
0
05 Mar 2025
CLIP is Strong Enough to Fight Back: Test-time Counterattacks towards Zero-shot Adversarial Robustness of CLIP
Songlong Xing
Zhengyu Zhao
N. Sebe
AAML
165
2
0
05 Mar 2025
Towards Effective and Sparse Adversarial Attack on Spiking Neural Networks via Breaking Invisible Surrogate Gradients
Li Lun
Kunyu Feng
Qinglong Ni
Ling Liang
Yuan Wang
Ying Li
Dunshan Yu
Xiaoxin Cui
AAML
117
0
0
05 Mar 2025
One Stone, Two Birds: Enhancing Adversarial Defense Through the Lens of Distributional Discrepancy
Jiacheng Zhang
Benjamin I. P. Rubinstein
Jing Zhang
Feng Liu
133
0
0
04 Mar 2025
Adversarial Agents: Black-Box Evasion Attacks with Reinforcement Learning
Kyle Domico
Jean-Charles Noirot Ferrand
Ryan Sheatsley
Eric Pauley
Josiah Hanna
Patrick McDaniel
AAML
122
1
0
03 Mar 2025
AutoAdvExBench: Benchmarking autonomous exploitation of adversarial example defenses
Nicholas Carlini
Javier Rando
Edoardo Debenedetti
Milad Nasr
F. Tramèr
AAML
ELM
92
3
0
03 Mar 2025
Exploiting Vulnerabilities in Speech Translation Systems through Targeted Adversarial Attacks
Chang-rui Liu
Haolin Wu
Xi Yang
Kui Zhang
Cong Wu
Weinan Zhang
Nenghai Yu
Tianwei Zhang
Qing Guo
Jie Zhang
AAML
66
0
0
02 Mar 2025
TAET: Two-Stage Adversarial Equalization Training on Long-Tailed Distributions
Wang YuHang
Junkang Guo
Aolei Liu
Kaihao Wang
Zaitong Wu
Zhenyu Liu
Wenfei Yin
Jian Liu
AAML
101
0
0
02 Mar 2025
A Guide to Failure in Machine Learning: Reliability and Robustness from Foundations to Practice
Eric Heim
Oren Wright
David Shriver
OOD
FaML
130
0
0
01 Mar 2025
Adversarial Attacks on Event-Based Pedestrian Detectors: A Physical Approach
Guixu Lin
Muyao Niu
Qingtian Zhu
Zhengwei Yin
Zhuoxiao Li
Shengfeng He
Yinqiang Zheng
AAML
80
0
0
01 Mar 2025
A Survey of Adversarial Defenses in Vision-based Systems: Categorization, Methods and Challenges
Nandish Chattopadhyay
Abdul Basit
B. Ouni
Muhammad Shafique
AAML
89
0
0
01 Mar 2025
Data-free Universal Adversarial Perturbation with Pseudo-semantic Prior
Chanhui Lee
Yeonghwan Song
Jeany Son
AAML
433
0
0
28 Feb 2025
Exploring the Impact of Temperature Scaling in Softmax for Classification and Adversarial Robustness
Hao Xuan
Bokai Yang
Xingyu Li
AAML
97
4
0
28 Feb 2025
À la recherche du sens perdu: your favourite LLM might have more to say than you can understand
K. O. T. Erziev
94
0
0
28 Feb 2025
Adversarial Prompt Evaluation: Systematic Benchmarking of Guardrails Against Prompt Input Attacks on LLMs
Giulio Zizzo
Giandomenico Cornacchia
Kieran Fraser
Muhammad Zaid Hameed
Ambrish Rawat
Beat Buesser
Mark Purcell
Pin-Yu Chen
P. Sattigeri
Kush R. Varshney
AAML
118
5
0
24 Feb 2025
EigenShield: Causal Subspace Filtering via Random Matrix Theory for Adversarially Robust Vision-Language Models
Nastaran Darabi
Devashri Naik
Sina Tayebati
Dinithi Jayasuriya
Ranganath Krishnan
A. R. Trivedi
AAML
165
0
0
24 Feb 2025
Single-pass Detection of Jailbreaking Input in Large Language Models
Leyla Naz Candogan
Yongtao Wu
Elias Abad Rocamora
Grigorios G. Chrysos
Volkan Cevher
AAML
118
0
0
24 Feb 2025
Unified Prompt Attack Against Text-to-Image Generation Models
Duo Peng
Qiuhong Ke
Mark He Huang
Ping Hu
Jing Liu
91
1
0
23 Feb 2025
SEA: Shareable and Explainable Attribution for Query-based Black-box Attacks
Yue Gao
Ilia Shumailov
Kassem Fawaz
AAML
224
0
0
21 Feb 2025
CyberSentinel: An Emergent Threat Detection System for AI Security
Krti Tallam
86
4
0
20 Feb 2025
Adversary-Aware DPO: Enhancing Safety Alignment in Vision Language Models via Adversarial Training
Fenghua Weng
Jian Lou
Jun Feng
Minlie Huang
Wenjie Wang
AAML
162
2
0
17 Feb 2025
PAR-AdvGAN: Improving Adversarial Attack Capability with Progressive Auto-Regression AdvGAN
Jiayu Zhang
Zhiyu Zhu
Xinyi Wang
Silin Liao
Zhibo Jin
Flora Salim
Huaming Chen
GAN
121
0
0
16 Feb 2025
Wasserstein distributional adversarial training for deep neural networks
Xingjian Bai
Guangyi He
Yifan Jiang
Jan Obloj
OOD
155
0
0
13 Feb 2025
Universal Adversarial Attack on Aligned Multimodal LLMs
Temurbek Rahmatullaev
Polina Druzhinina
Nikita Kurdiukov
Matvey Mikhalchuk
Andrey Kuznetsov
Anton Razzhigaev
AAML
223
0
0
11 Feb 2025
Adversarial Machine Learning: Attacks, Defenses, and Open Challenges
Pranav K Jha
AAML
94
0
0
08 Feb 2025
Democratic Training Against Universal Adversarial Perturbations
Bing-Jie Sun
Jun Sun
Wei Zhao
AAML
122
0
0
08 Feb 2025
Detecting APT Malware Command and Control over HTTP(S) Using Contextual Summaries
Almuthanna Alageel
Sergio Maffeis
Imperial College London
77
2
0
07 Feb 2025
Improving Adversarial Robustness via Phase and Amplitude-aware Prompting
Yibo Xu
Dawei Zhou
Decheng Liu
N. Wang
AAML
92
0
0
06 Feb 2025
Adversarial ML Problems Are Getting Harder to Solve and to Evaluate
Javier Rando
Jie Zhang
Nicholas Carlini
F. Tramèr
AAML
ELM
141
9
0
04 Feb 2025
Previous
1
2
3
4
5
6
...
79
80
81
Next