ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1706.06083
  4. Cited By
Towards Deep Learning Models Resistant to Adversarial Attacks
v1v2v3v4 (latest)

Towards Deep Learning Models Resistant to Adversarial Attacks

19 June 2017
Aleksander Madry
Aleksandar Makelov
Ludwig Schmidt
Dimitris Tsipras
Adrian Vladu
    SILMOOD
ArXiv (abs)PDFHTMLGithub (752★)

Papers citing "Towards Deep Learning Models Resistant to Adversarial Attacks"

50 / 6,612 papers shown
Title
Model-Free Adversarial Purification via Coarse-To-Fine Tensor Network Representation
Model-Free Adversarial Purification via Coarse-To-Fine Tensor Network Representation
Guang Lin
D. Nguyen
Zerui Tao
Konstantinos Slavakis
Toshihisa Tanaka
Qibin Zhao
AAML
110
1
0
25 Feb 2025
A stochastic smoothing framework for nonconvex-nonconcave min-sum-max problems with applications to Wasserstein distributionally robust optimization
A stochastic smoothing framework for nonconvex-nonconcave min-sum-max problems with applications to Wasserstein distributionally robust optimization
Wei Liu
Muhammad Khan
Gabriel Mancino-Ball
Yangyang Xu
82
1
0
24 Feb 2025
MACPruning: Dynamic Operation Pruning to Mitigate Side-Channel DNN Model Extraction
MACPruning: Dynamic Operation Pruning to Mitigate Side-Channel DNN Model Extraction
Ruyi Ding
Cheng Gongye
Davis Ranney
A. A. Ding
Yunsi Fei
AAML
112
0
0
24 Feb 2025
Interpreting Adversarial Attacks and Defences using Architectures with Enhanced Interpretability
Interpreting Adversarial Attacks and Defences using Architectures with Enhanced Interpretability
Akshay G Rao
Chandrashekhar Lakshminarayanan
Arun Rajkumar
AI4CEAAML
69
0
0
24 Feb 2025
Improving the Transferability of Adversarial Examples by Inverse Knowledge Distillation
Improving the Transferability of Adversarial Examples by Inverse Knowledge Distillation
Wenyuan Wu
Zheng Liu
Yong Chen
Chao Su
Dezhong Peng
Xu Wang
AAML
207
0
0
24 Feb 2025
Improved Diffusion-based Generative Model with Better Adversarial Robustness
Improved Diffusion-based Generative Model with Better Adversarial Robustness
Zekun Wang
Mingyang Yi
Shuchen Xue
Zhiyu Li
Ming Liu
Bing Qin
Zhi-Ming Ma
DiffM
116
0
0
24 Feb 2025
SMTFL: Secure Model Training to Untrusted Participants in Federated Learning
SMTFL: Secure Model Training to Untrusted Participants in Federated Learning
Zhihui Zhao
Xiaorong Dong
Yimo Ren
Jianhua Wang
Dan Yu
Hongsong Zhu
Yongle Chen
167
0
0
24 Feb 2025
EigenShield: Causal Subspace Filtering via Random Matrix Theory for Adversarially Robust Vision-Language Models
EigenShield: Causal Subspace Filtering via Random Matrix Theory for Adversarially Robust Vision-Language Models
Nastaran Darabi
Devashri Naik
Sina Tayebati
Dinithi Jayasuriya
Ranganath Krishnan
A. R. Trivedi
AAML
165
0
0
24 Feb 2025
Class-Conditional Neural Polarizer: A Lightweight and Effective Backdoor Defense by Purifying Poisoned Features
Class-Conditional Neural Polarizer: A Lightweight and Effective Backdoor Defense by Purifying Poisoned Features
Mingli Zhu
Shaokui Wei
Hongyuan Zha
Baoyuan Wu
AAML
123
0
0
23 Feb 2025
Can Indirect Prompt Injection Attacks Be Detected and Removed?
Can Indirect Prompt Injection Attacks Be Detected and Removed?
Yulin Chen
Haoran Li
Yuan Sui
Yufei He
Yue Liu
Yangqiu Song
Bryan Hooi
AAML
86
7
0
23 Feb 2025
Unified Prompt Attack Against Text-to-Image Generation Models
Unified Prompt Attack Against Text-to-Image Generation Models
Duo Peng
Qiuhong Ke
Mark He Huang
Ping Hu
Jing Liu
89
1
0
23 Feb 2025
Tracking the Copyright of Large Vision-Language Models through Parameter Learning Adversarial Images
Tracking the Copyright of Large Vision-Language Models through Parameter Learning Adversarial Images
Yubo Wang
Jianting Tang
Chaohu Liu
Linli Xu
AAML
189
1
0
23 Feb 2025
A generative approach to LLM harmfulness detection with special red flag tokens
A generative approach to LLM harmfulness detection with special red flag tokens
Sophie Xhonneux
David Dobre
Mehrnaz Mohfakhami
Leo Schwinn
Gauthier Gidel
184
2
0
22 Feb 2025
Tight Clusters Make Specialized Experts
Tight Clusters Make Specialized Experts
Stefan K. Nielsen
R. Teo
Laziz U. Abdullaev
Tan M. Nguyen
MoE
142
4
0
21 Feb 2025
Nearshore Underwater Target Detection Meets UAV-borne Hyperspectral Remote Sensing: A Novel Hybrid-level Contrastive Learning Framework and Benchmark Dataset
Nearshore Underwater Target Detection Meets UAV-borne Hyperspectral Remote Sensing: A Novel Hybrid-level Contrastive Learning Framework and Benchmark Dataset
Jiahao Qi
Chuanhong Zhou
Xingyue Liu
Chen Chen
Dehui Zhu
Kangcheng Bin
Ping Zhong
106
0
0
21 Feb 2025
SEA: Shareable and Explainable Attribution for Query-based Black-box Attacks
SEA: Shareable and Explainable Attribution for Query-based Black-box Attacks
Yue Gao
Ilia Shumailov
Kassem Fawaz
AAML
222
0
0
21 Feb 2025
Carefully Blending Adversarial Training, Purification, and Aggregation Improves Adversarial Robustness
Carefully Blending Adversarial Training, Purification, and Aggregation Improves Adversarial Robustness
Emanuele Ballarin
A. Ansuini
Luca Bortolussi
AAML
184
0
0
20 Feb 2025
CyberSentinel: An Emergent Threat Detection System for AI Security
CyberSentinel: An Emergent Threat Detection System for AI Security
Krti Tallam
86
4
0
20 Feb 2025
A Transfer Attack to Image Watermarks
A Transfer Attack to Image Watermarks
Yuepeng Hu
Zhengyuan Jiang
Moyang Guo
Neil Zhenqiang Gong
153
14
0
20 Feb 2025
Robust Optimization with Diffusion Models for Green Security
Robust Optimization with Diffusion Models for Green Security
Lingkai Kong
Haichuan Wang
Yuqi Pan
Cheol Woo Kim
Mingxiao Song
Alayna Nguyen
Tonghan Wang
Haifeng Xu
Milind Tambe
93
1
0
19 Feb 2025
Adversarially Robust CLIP Models Can Induce Better (Robust) Perceptual Metrics
Adversarially Robust CLIP Models Can Induce Better (Robust) Perceptual Metrics
Francesco Croce
Christian Schlarmann
Naman D. Singh
Matthias Hein
158
7
0
17 Feb 2025
Adversary-Aware DPO: Enhancing Safety Alignment in Vision Language Models via Adversarial Training
Adversary-Aware DPO: Enhancing Safety Alignment in Vision Language Models via Adversarial Training
Fenghua Weng
Jian Lou
Jun Feng
Minlie Huang
Wenjie Wang
AAML
160
2
0
17 Feb 2025
PAR-AdvGAN: Improving Adversarial Attack Capability with Progressive Auto-Regression AdvGAN
PAR-AdvGAN: Improving Adversarial Attack Capability with Progressive Auto-Regression AdvGAN
Jiayu Zhang
Zhiyu Zhu
Xinyi Wang
Silin Liao
Zhibo Jin
Flora Salim
Huaming Chen
GAN
117
0
0
16 Feb 2025
FaceSwapGuard: Safeguarding Facial Privacy from DeepFake Threats through Identity Obfuscation
FaceSwapGuard: Safeguarding Facial Privacy from DeepFake Threats through Identity Obfuscation
Li Wang
Zheng Li
Wei Wei
Shouling Ji
Shanqing Guo
PICVAAML
102
2
0
15 Feb 2025
Wasserstein distributional adversarial training for deep neural networks
Wasserstein distributional adversarial training for deep neural networks
Xingjian Bai
Guangyi He
Yifan Jiang
Jan Obloj
OOD
155
0
0
13 Feb 2025
DeepSeek on a Trip: Inducing Targeted Visual Hallucinations via Representation Vulnerabilities
DeepSeek on a Trip: Inducing Targeted Visual Hallucinations via Representation Vulnerabilities
Chashi Mahiul Islam
Samuel Jacob Chacko
Preston Horne
Xiuwen Liu
165
2
0
11 Feb 2025
Robust Watermarks Leak: Channel-Aware Feature Extraction Enables Adversarial Watermark Manipulation
Zhongjie Ba
Yize Zhang
Peng Cheng
Bin Gong
Xinyu Zhang
Qinglong Wang
Kui Ren
AAML
207
0
0
10 Feb 2025
Amnesia as a Catalyst for Enhancing Black Box Pixel Attacks in Image Classification and Object Detection
Amnesia as a Catalyst for Enhancing Black Box Pixel Attacks in Image Classification and Object Detection
Dongsu Song
Daehwa Ko
Jay Hoon Jung
AAML
102
0
0
10 Feb 2025
Effective Black-Box Multi-Faceted Attacks Breach Vision Large Language Model Guardrails
Effective Black-Box Multi-Faceted Attacks Breach Vision Large Language Model Guardrails
Yijun Yang
L. Wang
Xiao Yang
Lanqing Hong
Jun Zhu
AAML
75
0
0
09 Feb 2025
Sign-Symmetry Learning Rules are Robust Fine-Tuners
Sign-Symmetry Learning Rules are Robust Fine-Tuners
Aymene Berriche
Mehdi Zakaria Adjal
Riyadh Baghdadi
AAML
75
0
0
09 Feb 2025
Democratic Training Against Universal Adversarial Perturbations
Bing-Jie Sun
Jun Sun
Wei Zhao
AAML
122
0
0
08 Feb 2025
Adversarial Machine Learning: Attacks, Defenses, and Open Challenges
Adversarial Machine Learning: Attacks, Defenses, and Open Challenges
Pranav K Jha
AAML
94
0
0
08 Feb 2025
Federated Learning for Anomaly Detection in Energy Consumption Data: Assessing the Vulnerability to Adversarial Attacks
Federated Learning for Anomaly Detection in Energy Consumption Data: Assessing the Vulnerability to Adversarial Attacks
Yohannis Kifle Telila
Damitha Senevirathne
Dumindu Tissera
Apurva Narayan
Miriam A.M. Capretz
Katarina Grolinger
AAML
80
0
0
07 Feb 2025
Confidence Elicitation: A New Attack Vector for Large Language Models
Confidence Elicitation: A New Attack Vector for Large Language Models
Brian Formento
Chuan-Sheng Foo
See-Kiong Ng
AAML
266
0
0
07 Feb 2025
Improving Adversarial Robustness via Phase and Amplitude-aware Prompting
Improving Adversarial Robustness via Phase and Amplitude-aware Prompting
Yibo Xu
Dawei Zhou
Decheng Liu
N. Wang
AAML
89
0
0
06 Feb 2025
How vulnerable is my policy? Adversarial attacks on modern behavior cloning policies
How vulnerable is my policy? Adversarial attacks on modern behavior cloning policies
Basavasagar Patil
Akansha Kalra
Guanhong Tao
Daniel S. Brown
AAML
111
0
0
06 Feb 2025
Rotation-Adaptive Point Cloud Domain Generalization via Intricate Orientation Learning
Rotation-Adaptive Point Cloud Domain Generalization via Intricate Orientation Learning
Bangzhen Liu
Chenxi Zheng
Xuemiao Xu
Cheng Xu
Huaidong Zhang
Shengfeng He
3DPC
72
1
0
04 Feb 2025
Achievable distributional robustness when the robust risk is only partially identified
Achievable distributional robustness when the robust risk is only partially identified
Julia Kostin
Nicola Gnecco
Fanny Yang
153
3
0
04 Feb 2025
INTACT: Inducing Noise Tolerance through Adversarial Curriculum Training for LiDAR-based Safety-Critical Perception and Autonomy
INTACT: Inducing Noise Tolerance through Adversarial Curriculum Training for LiDAR-based Safety-Critical Perception and Autonomy
Nastaran Darabi
Divake Kumar
Sina Tayebati
A. R. Trivedi
AAML
139
0
0
04 Feb 2025
Analytical Lyapunov Function Discovery: An RL-based Generative Approach
Analytical Lyapunov Function Discovery: An RL-based Generative Approach
Haohan Zou
Jie Feng
Hao Zhao
Yuanyuan Shi
177
0
0
04 Feb 2025
Adversarial ML Problems Are Getting Harder to Solve and to Evaluate
Adversarial ML Problems Are Getting Harder to Solve and to Evaluate
Javier Rando
Jie Zhang
Nicholas Carlini
F. Tramèr
AAMLELM
139
9
0
04 Feb 2025
CoRPA: Adversarial Image Generation for Chest X-rays Using Concept Vector Perturbations and Generative Models
CoRPA: Adversarial Image Generation for Chest X-rays Using Concept Vector Perturbations and Generative Models
Amy Rafferty
Rishi Ramaesh
Ajitha Rajan
MedImAAML
147
0
0
04 Feb 2025
MoireDB: Formula-generated Interference-fringe Image Dataset
MoireDB: Formula-generated Interference-fringe Image Dataset
Yuto Matsuo
Ryo Hayamizu
Hirokatsu Kataoka
Akio Nakamura
74
0
0
03 Feb 2025
Robust-LLaVA: On the Effectiveness of Large-Scale Robust Image Encoders for Multi-modal Large Language Models
Robust-LLaVA: On the Effectiveness of Large-Scale Robust Image Encoders for Multi-modal Large Language Models
H. Malik
Fahad Shamshad
Muzammal Naseer
Karthik Nandakumar
Fahad Shahbaz Khan
Salman Khan
AAMLMLLMVLM
137
1
0
03 Feb 2025
Towards Robust Multimodal Large Language Models Against Jailbreak Attacks
Towards Robust Multimodal Large Language Models Against Jailbreak Attacks
Ziyi Yin
Yuanpu Cao
Han Liu
Ting Wang
Jinghui Chen
Fenhlong Ma
AAML
97
1
0
02 Feb 2025
Imitation Game for Adversarial Disillusion with Multimodal Generative Chain-of-Thought Role-Play
Imitation Game for Adversarial Disillusion with Multimodal Generative Chain-of-Thought Role-Play
Ching-Chun Chang
Fan-Yun Chen
Shih-Hong Gu
Kai Gao
Hanrui Wang
Isao Echizen
AAML
512
0
0
31 Jan 2025
Trading Inference-Time Compute for Adversarial Robustness
Trading Inference-Time Compute for Adversarial Robustness
Wojciech Zaremba
Evgenia Nitishinskaya
Boaz Barak
Stephanie Lin
Sam Toyer
...
Rachel Dias
Eric Wallace
Kai Y. Xiao
Johannes Heidecke
Amelia Glaese
LRMAAML
167
26
0
31 Jan 2025
Scanning Trojaned Models Using Out-of-Distribution Samples
Scanning Trojaned Models Using Out-of-Distribution Samples
Hossein Mirzaei
Ali Ansari
Bahar Dibaei Nia
Mojtaba Nafez
Moein Madadi
...
Kian Shamsaie
Mahdi Hajialilue
Jafar Habibi
Mohammad Sabokrou
M. Rohban
OODD
143
3
0
28 Jan 2025
Killing it with Zero-Shot: Adversarially Robust Novelty Detection
Hossein Mirzaei
Mohammad Jafari
Hamid Reza Dehbashi
Zeinab Sadat Taghavi
Mohammad Sabokrou
M. Rohban
119
2
0
28 Jan 2025
Improving Factuality in Large Language Models via Decoding-Time Hallucinatory and Truthful Comparators
Improving Factuality in Large Language Models via Decoding-Time Hallucinatory and Truthful Comparators
Dingkang Yang
Dongling Xiao
Jinjie Wei
Mingcheng Li
Zhaoyu Chen
Ke Li
Li Zhang
HILM
167
6
0
28 Jan 2025
Previous
123...789...131132133
Next