581

Lipschitz-Margin Training: Scalable Certification of Perturbation Invariance for Deep Neural Networks

Abstract

High sensitivity of neural networks against malicious perturbations on inputs causes security concerns. To take a steady step towards robust classifiers, we aim to create neural network models provably defended from perturbations. Prior certification work requires strong assumptions on network structures and massive computational costs, and thus, their applications are limited. Based on the relationship between the Lipschitz constants and prediction margins, we present a computationally efficient calculation technique that lower-bounds the size of adversarial perturbations that can deceive networks, and that is widely applicable to various complicated networks. Moreover, we propose an efficient training procedure, which robustifies networks and significantly improves the provably guarded areas around data points. In experimental evaluations, our method showed its ability to provide a non-trivial guarantee and improve robustness for even large networks.

View on arXiv
Comments on this paper