Distinction Maximization Loss: Fast, Scalable, Turnkey, and Native
Neural Networks Out-of-Distribution Detection simply by Replacing the SoftMax
Loss
IEEE International Joint Conference on Neural Network (IJCNN), 2019
- OODDUQCV
Abstract
Recently, many methods to reduce neural networks uncertainty have been proposed. However, most of the techniques used in these solutions usually present severe drawbacks. In this paper, we argue that neural networks low out-of-distribution detection performance is mainly due to the SoftMax loss anisotropy. Therefore, we built an isotropic loss to reduce neural networks uncertainty in a fast, scalable, turnkey, and native approach. Our experiments showed that our proposal overcomes ODIN typically by a large margin while producing usually competitive results against state-of-the-art Mahalanobis method while avoiding their limitations.
View on arXivComments on this paper
