Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1706.06083
Cited By
v1
v2
v3
v4 (latest)
Towards Deep Learning Models Resistant to Adversarial Attacks
19 June 2017
Aleksander Madry
Aleksandar Makelov
Ludwig Schmidt
Dimitris Tsipras
Adrian Vladu
SILM
OOD
Re-assign community
ArXiv (abs)
PDF
HTML
Github (752★)
Papers citing
"Towards Deep Learning Models Resistant to Adversarial Attacks"
50 / 6,612 papers shown
Title
Noninvasive precision modulation of high-level neural population activity via natural vision perturbations
Guy Gaziv
Sarah Goulding
Ani Ayvazian-Hancock
Yoon Bai
J. DiCarlo
34
0
0
05 Jun 2025
Identifying and Understanding Cross-Class Features in Adversarial Training
Zeming Wei
Yiwen Guo
Yisen Wang
AAML
102
0
0
05 Jun 2025
Fool the Stoplight: Realistic Adversarial Patch Attacks on Traffic Light Detectors
Svetlana Pavlitska
Jamie Robb
Nikolai Polley
Melih Yazgan
Johann Marius Zöllner
AAML
107
0
0
05 Jun 2025
RAID: A Dataset for Testing the Adversarial Robustness of AI-Generated Image Detectors
Hicham Eddoubi
Jonas Ricker
Federico Cocchi
Lorenzo Baraldi
Angelo Sotgiu
...
Marcella Cornia
Lorenzo Baraldi
Asja Fischer
Rita Cucchiara
Battista Biggio
AAML
147
0
0
04 Jun 2025
MUC-G4: Minimal Unsat Core-Guided Incremental Verification for Deep Neural Network Compression
Jingyang Li
Guoqiang Li
24
0
0
03 Jun 2025
Privacy Leaks by Adversaries: Adversarial Iterations for Membership Inference Attack
Jing Xue
Zhishen Sun
Haishan Ye
Luo Luo
Xiangyu Chang
Ivor Tsang
Guang Dai
MIACV
MIALM
64
0
0
03 Jun 2025
Dynamic Epsilon Scheduling: A Multi-Factor Adaptive Perturbation Budget for Adversarial Training
Alan Mitkiy
James Smith
Hana Satou
Hiroshi Tanaka
Emily Johnson
F Monkey
AAML
15
0
0
03 Jun 2025
On the Robustness of Tabular Foundation Models: Test-Time Attacks and In-Context Defenses
Mohamed Djilani
Thibault Simonetto
Karim Tit
Florian Tambon
Paul Récamier
Salah Ghamizi
Maxime Cordy
Mike Papadakis
AAML
50
0
0
03 Jun 2025
Attacking Attention of Foundation Models Disrupts Downstream Tasks
Hondamunige Prasanna Silva
Federico Becattini
Lorenzo Seidenari
AAML
27
0
0
03 Jun 2025
Urban Visibility Hotspots: Quantifying Building Vertex Visibility from Connected Vehicle Trajectories using Spatial Indexing
Artur Grigorev
Adriana-Simona Mihaita
38
0
0
03 Jun 2025
A Trustworthiness-based Metaphysics of Artificial Intelligence Systems
Andrea Ferrario
36
0
0
03 Jun 2025
Silence is Golden: Leveraging Adversarial Examples to Nullify Audio Control in LDM-based Talking-Head Generation
Yuan Gan
Jiaxu Miao
Yunze Wang
Yi Yang
AAML
DiffM
49
0
0
02 Jun 2025
Align is not Enough: Multimodal Universal Jailbreak Attack against Multimodal Large Language Models
Youze Wang
Wenbo Hu
Yinpeng Dong
Jing Liu
Hanwang Zhang
Richang Hong
57
2
0
02 Jun 2025
Unlearning's Blind Spots: Over-Unlearning and Prototypical Relearning Attack
SeungBum Ha
Saerom Park
Sung Whan Yoon
MU
CLL
60
0
1
02 Jun 2025
Enhancing Diffusion-based Unrestricted Adversarial Attacks via Adversary Preferences Alignment
Kaixun Jiang
Zhaoyu Chen
Haijing Guo
Jinglun Li
Jiyuan Fu
Pinxue Guo
Hao Tang
Bo Li
Wenqiang Zhang
DiffM
AAML
84
0
0
02 Jun 2025
Comprehensive Vulnerability Analysis is Necessary for Trustworthy LLM-MAS
Pengfei He
Yue Xing
Shen Dong
Juanhui Li
Zhenwei Dai
...
Hui Liu
Han Xu
Zhen Xiang
Charu C. Aggarwal
Hui Liu
LLMAG
84
0
0
02 Jun 2025
Fighting Fire with Fire (F3): A Training-free and Efficient Visual Adversarial Example Purification Method in LVLMs
Yudong Zhang
Ruobing Xie
Yiqing Huang
Jiansheng Chen
Xingwu Sun
Zhanhui Kang
Di Wang
Yu Wang
AAML
49
0
0
01 Jun 2025
No Soundness in the Real World: On the Challenges of the Verification of Deployed Neural Networks
Attila Szász
Balázs Bánhelyi
Márk Jelasity
25
0
0
01 Jun 2025
On the Stability of Graph Convolutional Neural Networks: A Probabilistic Perspective
Mihai Cucuringu
Henry Kenlay
Li Zhang
Ning Zhang
Xiaowen Dong
GNN
AAML
66
0
0
01 Jun 2025
The Security Threat of Compressed Projectors in Large Vision-Language Models
Yudong Zhang
Ruobing Xie
Xingwu Sun
Jiansheng Chen
Zhanhui Kang
Di Wang
Yu Wang
16
0
0
31 May 2025
PatchDEMUX: A Certifiably Robust Framework for Multi-label Classifiers Against Adversarial Patches
Dennis Jacob
Chong Xiang
Prateek Mittal
AAML
15
0
0
30 May 2025
A Red Teaming Roadmap Towards System-Level Safety
Zifan Wang
Christina Q. Knight
Jeremy Kritz
Willow Primack
Julian Michael
AAML
45
0
0
30 May 2025
Black-box Adversarial Attacks on CNN-based SLAM Algorithms
M. Gkeka
Bowen Sun
Evgenia Smirni
C. Antonopoulos
S. Lalis
Nikolaos Bellas
AAML
27
0
0
30 May 2025
Adversarial Preference Learning for Robust LLM Alignment
Yuanfu Wang
Pengyu Wang
Chenyang Xi
Bo Tang
Junyi Zhu
...
Keming Mao
Zhiyu Li
Feiyu Xiong
Jie Hu
Mingchuan Yang
AAML
24
0
0
30 May 2025
Adversarial Semantic and Label Perturbation Attack for Pedestrian Attribute Recognition
Weizhe Kong
Xiao Wang
Ruichong Gao
Chenglong Li
Yu Zhang
Xing Yang
Yaowei Wang
Jin Tang
AAML
64
0
0
29 May 2025
TRAP: Targeted Redirecting of Agentic Preferences
Hangoo Kang
Jehyeok Yeon
Gagandeep Singh
AAML
72
0
0
29 May 2025
Understanding Adversarial Training with Energy-based Models
Mujtaba Hussain Mirza
Maria Rosaria Briglia
Filippo Bartolucci
Senad Beadini
G. Lisanti
I. Masi
AAML
57
0
0
28 May 2025
Spa-VLM: Stealthy Poisoning Attacks on RAG-based VLM
Lei Yu
Yechao Zhang
Ziqi Zhou
Yang Wu
Wei Wan
Minghui Li
Shengshan Hu
Pei Xiaobing
Jing Wang
AAML
27
0
0
28 May 2025
How Do Diffusion Models Improve Adversarial Robustness?
Liu Yuezhang
Xue-Xin Wei
296
0
0
28 May 2025
Distributionally Robust Wireless Semantic Communication with Large AI Models
Long Tan Le
Senura Hansaja Wanasekara
Zerun Niu
Yansong Shi
Nguyen Tran
...
Walid Saad
Dusit Niyato
Zhu Han
Choong Seon Hong
H. V. Poor
20
0
0
28 May 2025
Preventing Adversarial AI Attacks Against Autonomous Situational Awareness: A Maritime Case Study
Mathew J. Walter
Aaron Barrett
Kimberly Tam
AAML
32
1
0
27 May 2025
One-Time Soft Alignment Enables Resilient Learning without Weight Transport
Jeonghwan Cheon
Jaehyuk Bae
Se-Bum Paik
ODL
57
1
0
27 May 2025
NatADiff: Adversarial Boundary Guidance for Natural Adversarial Diffusion
Max Collins
Jordan Vice
T. French
Ajmal Mian
DiffM
48
0
0
27 May 2025
Breaking Dataset Boundaries: Class-Agnostic Targeted Adversarial Attacks
Taïga Gonçalves
Tomo Miyazaki
S. Omachi
OOD
AAML
81
0
0
27 May 2025
VideoMarkBench: Benchmarking Robustness of Video Watermarking
Zhengyuan Jiang
Moyang Guo
Kecen Li
Yuepeng Hu
Yupu Wang
Zhicong Huang
Cheng Hong
Neil Zhenqiang Gong
AAML
28
0
0
27 May 2025
Adversarial Attacks against Closed-Source MLLMs via Feature Optimal Alignment
Xiaojun Jia
Sensen Gao
Simeng Qin
Tianyu Pang
C. Du
Yihao Huang
Xinfeng Li
Yiming Li
Bo Li
Yang Liu
AAML
44
0
0
27 May 2025
Model Stitching by Functional Latent Alignment
Ioannis Athanasiadis
Anmar Karmush
Michael Felsberg
56
0
0
26 May 2025
Novel Loss-Enhanced Universal Adversarial Patches for Sustainable Speaker Privacy
Elvir Karimov
Alexander Varlamov
Danil Ivanov
Dmitrii Korzh
Oleg Y. Rogov
AAML
34
0
0
26 May 2025
Are Time-Series Foundation Models Deployment-Ready? A Systematic Study of Adversarial Robustness Across Domains
Jiawen Zhang
Zhenwei Zhang
Shun Zheng
Xumeng Wen
Jia Li
Jiang Bian
AI4TS
AAML
163
0
0
26 May 2025
Attention! You Vision Language Model Could Be Maliciously Manipulated
Xiaosen Wang
Shaokang Wang
Zhijin Ge
Yuyang Luo
Shudong Zhang
AAML
VLM
39
0
0
26 May 2025
Diagnosing and Mitigating Modality Interference in Multimodal Large Language Models
Rui Cai
Bangzheng Li
Xiaofei Wen
Muhao Chen
Zhe Zhao
24
0
0
26 May 2025
TESSER: Transfer-Enhancing Adversarial Attacks from Vision Transformers via Spectral and Semantic Regularization
Amira Guesmi
B. Ouni
Muhammad Shafique
AAML
233
0
0
26 May 2025
JailBound: Jailbreaking Internal Safety Boundaries of Vision-Language Models
Jiaxin Song
Yixu Wang
Jie Li
Rui Yu
Yan Teng
Xingjun Ma
Yingchun Wang
AAML
60
0
0
26 May 2025
Benign-to-Toxic Jailbreaking: Inducing Harmful Responses from Harmless Prompts
H. Kim
Minbeom Kim
Wonjun Lee
Kihyun Kim
Changick Kim
32
0
0
26 May 2025
Continuous-Time Analysis of Heavy Ball Momentum in Min-Max Games
Yi-Hu Feng
Kaito Fujii
Stratis Skoulakis
Xiao Wang
Volkan Cevher
96
0
0
26 May 2025
One Surrogate to Fool Them All: Universal, Transferable, and Targeted Adversarial Attacks with CLIP
Binyan Xu
Xilin Dai
Di Tang
Kehuan Zhang
AAML
22
0
0
26 May 2025
RADEP: A Resilient Adaptive Defense Framework Against Model Extraction Attacks
Amit Chakraborty
Sayyed Farid Ahamed
Sandip Roy
S. Banerjee
Kevin Choi
A. Rahman
Alison Hu
Edward Bowen
Sachin Shetty
AAML
55
0
0
25 May 2025
Ignition Phase : Standard Training for Fast Adversarial Robustness
Wang Yu-Hang
Liu ying
Fang liang
Wang Xuelin
Junkang Guo
Shiwei Li
Lei Gao
Jian Liu
Wenfei Yin
AAML
14
0
0
25 May 2025
LORE: Lagrangian-Optimized Robust Embeddings for Visual Encoders
Borna Khodabandeh
Amirabbas Afzali
Amirhossein Afsharrad
Seyed Shahabeddin Mousavi
Sanjay Lall
Sajjad Amini
Seyed-Mohsen Moosavi-Dezfooli
AAML
38
0
0
24 May 2025
Adversarial Robustness of Nonparametric Regression
Parsa Moradi
Hanzaleh Akabrinodehi
M. Maddah-ali
AAML
76
0
0
23 May 2025
Previous
1
2
3
4
5
...
131
132
133
Next