ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2103.07364
64
22

A Unified Game-Theoretic Interpretation of Adversarial Robustness

12 March 2021
Jie Ren
Die Zhang
Yisen Wang
Lu Chen
Zhanpeng Zhou
Yiting Chen
Xu Cheng
Xin Eric Wang
Meng Zhou
Jie Shi
Quanshi Zhang
    AAML
ArXivPDFHTML
Abstract

This paper provides a unified view to explain different adversarial attacks and defense methods, i.e. the view of multi-order interactions between input variables of DNNs. Based on the multi-order interaction, we discover that adversarial attacks mainly affect high-order interactions to fool the DNN. Furthermore, we find that the robustness of adversarially trained DNNs comes from category-specific low-order interactions. Our findings provide a potential method to unify adversarial perturbations and robustness, which can explain the existing defense methods in a principle way. Besides, our findings also make a revision of previous inaccurate understanding of the shape bias of adversarially learned features.

View on arXiv
Comments on this paper