ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2003.06974
79
13

Toward Adversarial Robustness via Semi-supervised Robust Training

16 March 2020
Yiming Li
Baoyuan Wu
Yan Feng
Yanbo Fan
Yong Jiang
Zhifeng Li
Shutao Xia
    AAML
ArXivPDFHTML
Abstract

Adversarial examples have been shown to be the severe threat to deep neural networks (DNNs). One of the most effective adversarial defense methods is adversarial training (AT) through minimizing the adversarial risk RadvR_{adv}Radv​, which encourages both the benign example xxx and its adversarially perturbed neighborhoods within the ℓp\ell_{p}ℓp​-ball to be predicted as the ground-truth label. In this work, we propose a novel defense method, the robust training (RT), by jointly minimizing two separated risks (RstandR_{stand}Rstand​ and RrobR_{rob}Rrob​), which is with respect to the benign example and its neighborhoods respectively. The motivation is to explicitly and jointly enhance the accuracy and the adversarial robustness. We prove that RadvR_{adv}Radv​ is upper-bounded by Rstand+RrobR_{stand} + R_{rob}Rstand​+Rrob​, which implies that RT has similar effect as AT. Intuitively, minimizing the standard risk enforces the benign example to be correctly predicted, and the robust risk minimization encourages the predictions of the neighbor examples to be consistent with the prediction of the benign example. Besides, since RrobR_{rob}Rrob​ is independent of the ground-truth label, RT is naturally extended to the semi-supervised mode (i.e.i.e.i.e., SRT), to further enhance the adversarial robustness. Moreover, we extend the ℓp\ell_{p}ℓp​-bounded neighborhood to a general case, which covers different types of perturbations, such as the pixel-wise (i.e.i.e.i.e., x+δx + \deltax+δ) or the spatial perturbation (i.e.i.e.i.e., AX+b AX + bAX+b). Extensive experiments on benchmark datasets not only verify the superiority of the proposed SRT method to state-of-the-art methods for defensing pixel-wise or spatial perturbations separately, but also demonstrate its robustness to both perturbations simultaneously. The code for reproducing main results is available at \url{https://github.com/THUYimingLi/Semi-supervised_Robust_Training}.

View on arXiv
Comments on this paper