Exploiting Verified Neural Networks via Floating Point Numerical Error
Sensors Applications Symposium (SA), 2020
- AAML
Abstract
We show how to construct adversarial examples for neural networks with exactly verified robustness against -bounded input perturbations by exploiting floating point error. We argue that any exact verification of real-valued neural networks must accurately model the implementation details of any floating point arithmetic used during inference or verification.
View on arXivComments on this paper
