Adversarial samples are maliciously created inputs that lead a machine learning classifier to produce incorrect output labels. An adversarial sample is often generated by adding adversarial noise (AN) to a normal test sample. Recent literature has tried to analyze and harden learning-based classifiers under such AN. However, previous studies are mostly empirical and provide little understanding of why many learning-based classifiers, including deep neural networks (DNNs), are vulnerable to AN. To fill this gap, we propose a theoretical framework using the notation of topological spaces to uncover such reasons. The central idea of our work is that for a certain classification task, the robustness of a classifier against AN is decided by both and its oracle (such as human annotators of that specific task). This motivates us to formulate a formal definition of "strong-robustness" that describes when a classifier is always robust against AN according to its . The second key piece of our framework is the decomposition of , in which , includes feature learning operations and includes relatively simple decision functions for the classification. We theoretically prove that is strong-robust against AN if and only if a special topological relationship exists between the two feature spaces defined by and . Theorems of our framework provide two important insights: (1) The strong-robustness of is fully determined by its , not . (2) Extra irrelevant features ruin the strong-robustness of . Empirically we find that the Siamese architecture can intuitively help DNN models approach the desired topological relationship for strong-robustness, which in turn effectively improves its robustness against AN.
View on arXiv