ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1912.12170
11
0

Mitigating large adversarial perturbations on X-MAS (X minus Moving Averaged Samples)

19 December 2019
Woohyung Chun
Sung-Min Hong
Junho Huh
Inyup Kang
    AAML
ArXivPDFHTML
Abstract

We propose the scheme that mitigates the adversarial perturbation ϵ\epsilonϵ on the adversarial example XadvX_{adv}Xadv​ (=== XXX ±\pm± ϵ\epsilonϵ, XXX is a benign sample) by subtracting the estimated perturbation ϵ^\hat{\epsilon}ϵ^ from XXX +++ ϵ\epsilonϵ and adding ϵ^\hat{\epsilon}ϵ^ to XXX −-− ϵ\epsilonϵ. The estimated perturbation ϵ^\hat{\epsilon}ϵ^ comes from the difference between XadvX_{adv}Xadv​ and its moving-averaged outcome Wavg∗XadvW_{avg}*X_{adv}Wavg​∗Xadv​ where WavgW_{avg}Wavg​ is N×NN \times NN×N moving average kernel that all the coefficients are one. Usually, the adjacent samples of an image are close to each other such that we can let XXX ≈\approx≈ Wavg∗XW_{avg}*XWavg​∗X (naming this relation after X-MAS[X minus Moving Averaged Samples]). By doing that, we can make the estimated perturbation ϵ^\hat{\epsilon}ϵ^ falls within the range of ϵ\epsilonϵ. The scheme is also extended to do the multi-level mitigation by configuring the mitigated adversarial example XadvX_{adv}Xadv​ ±\pm± ϵ^\hat{\epsilon}ϵ^ as a new adversarial example to be mitigated. The multi-level mitigation gets XadvX_{adv}Xadv​ closer to XXX with a smaller (i.e. mitigated) perturbation than original unmitigated perturbation by setting the moving averaged adversarial sample Wavg∗XadvW_{avg} * X_{adv}Wavg​∗Xadv​ (which has the smaller perturbation than XadvX_{adv}Xadv​ if XXX ≈\approx≈ Wavg∗XW_{avg}*XWavg​∗X) as the boundary condition that the multi-level mitigation cannot cross over (i.e. decreasing ϵ\epsilonϵ cannot go below and increasing ϵ\epsilonϵ cannot go beyond). With the multi-level mitigation, we can get high prediction accuracies even in the adversarial example having a large perturbation (i.e. ϵ\epsilonϵ >>> 161616). The proposed scheme is evaluated with adversarial examples crafted by the FGSM (Fast Gradient Sign Method) based attacks on ResNet-50 trained with ImageNet dataset.

View on arXiv
Comments on this paper