Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2201.13329
Cited By
Can Adversarial Training Be Manipulated By Non-Robust Features?
31 January 2022
Lue Tao
Lei Feng
Hongxin Wei
Jinfeng Yi
Sheng-Jun Huang
Songcan Chen
AAML
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Can Adversarial Training Be Manipulated By Non-Robust Features?"
16 / 16 papers shown
Title
A Survey on Adversarial Machine Learning for Code Data: Realistic Threats, Countermeasures, and Interpretations
Yulong Yang
Haoran Fan
Chenhao Lin
Qian Li
Zhengyu Zhao
Chao Shen
Xiaohong Guan
AAML
48
0
0
12 Nov 2024
Deferred Poisoning: Making the Model More Vulnerable via Hessian Singularization
Yuhao He
Jinyu Tian
Xianwei Zheng
Li Dong
Yuanman Li
L. Zhang
AAML
28
0
0
06 Nov 2024
Adversarial Training: A Survey
Mengnan Zhao
Lihe Zhang
Jingwen Ye
Huchuan Lu
Baocai Yin
Xinchao Wang
AAML
28
1
0
19 Oct 2024
Toward Availability Attacks in 3D Point Clouds
Yifan Zhu
Yibo Miao
Yinpeng Dong
Xiao-Shan Gao
3DPC
AAML
51
3
0
26 Jun 2024
Nonlinear Transformations Against Unlearnable Datasets
T. Hapuarachchi
Jing Lin
Kaiqi Xiong
Mohamed Rahouti
Gitte Ost
28
1
0
05 Jun 2024
Theoretical Understanding of Learning from Adversarial Perturbations
Soichiro Kumano
Hiroshi Kera
Toshihiko Yamasaki
AAML
36
1
0
16 Feb 2024
Efficient Availability Attacks against Supervised and Contrastive Learning Simultaneously
Yihan Wang
Yifan Zhu
Xiao-Shan Gao
AAML
33
6
0
06 Feb 2024
Data-Dependent Stability Analysis of Adversarial Training
Yihan Wang
Shuang Liu
Xiao-Shan Gao
36
3
0
06 Jan 2024
Detection and Defense of Unlearnable Examples
Yifan Zhu
Lijia Yu
Xiao-Shan Gao
AAML
24
7
0
14 Dec 2023
Adversarial Examples Are Not Real Features
Ang Li
Yifei Wang
Yiwen Guo
Yisen Wang
20
11
0
29 Oct 2023
What Distributions are Robust to Indiscriminate Poisoning Attacks for Linear Learners?
Fnu Suya
X. Zhang
Yuan Tian
David E. Evans
OOD
AAML
26
2
0
03 Jul 2023
Average of Pruning: Improving Performance and Stability of Out-of-Distribution Detection
Zhen Cheng
Fei Zhu
Xu-Yao Zhang
Cheng-Lin Liu
MoMe
OODD
40
11
0
02 Mar 2023
Unlearnable Examples: Making Personal Data Unexploitable
Hanxun Huang
Xingjun Ma
S. Erfani
James Bailey
Yisen Wang
MIACV
156
190
0
13 Jan 2021
Unadversarial Examples: Designing Objects for Robust Vision
Hadi Salman
Andrew Ilyas
Logan Engstrom
Sai H. Vemprala
A. Madry
Ashish Kapoor
WIGM
65
59
0
22 Dec 2020
RobustBench: a standardized adversarial robustness benchmark
Francesco Croce
Maksym Andriushchenko
Vikash Sehwag
Edoardo Debenedetti
Nicolas Flammarion
M. Chiang
Prateek Mittal
Matthias Hein
VLM
234
678
0
19 Oct 2020
Adversarial Machine Learning at Scale
Alexey Kurakin
Ian Goodfellow
Samy Bengio
AAML
293
3,112
0
04 Nov 2016
1