ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1906.02931
17
16

Inductive Bias of Gradient Descent based Adversarial Training on Separable Data

7 June 2019
Yan Li
Ethan X. Fang
Huan Xu
T. Zhao
ArXivPDFHTML
Abstract

Adversarial training is a principled approach for training robust neural networks. Despite of tremendous successes in practice, its theoretical properties still remain largely unexplored. In this paper, we provide new theoretical insights of gradient descent based adversarial training by studying its computational properties, specifically on its inductive bias. We take the binary classification task on linearly separable data as an illustrative example, where the loss asymptotically attains its infimum as the parameter diverges to infinity along certain directions. Specifically, we show that when the adversarial perturbation during training has bounded ℓ2\ell_2ℓ2​-norm, the classifier learned by gradient descent based adversarial training converges in direction to the maximum ℓ2\ell_2ℓ2​-norm margin classifier at the rate of O~(1/T)\tilde{\mathcal{O}}(1/\sqrt{T})O~(1/T​), significantly faster than the rate O(1/log⁡T)\mathcal{O}(1/\log T)O(1/logT) of training with clean data. In addition, when the adversarial perturbation during training has bounded ℓq\ell_qℓq​-norm for some q≥1q\ge 1q≥1, the resulting classifier converges in direction to a maximum mixed-norm margin classifier, which has a natural interpretation of robustness, as being the maximum ℓ2\ell_2ℓ2​-norm margin classifier under worst-case ℓq\ell_qℓq​-norm perturbation to the data. Our findings provide theoretical backups for adversarial training that it indeed promotes robustness against adversarial perturbation.

View on arXiv
Comments on this paper