25
0
v1v2 (latest)

LoRA as a Flexible Framework for Securing Large Vision Systems

Main:4 Pages
2 Figures
Bibliography:3 Pages
4 Tables
Appendix:4 Pages
Abstract

Adversarial attacks have emerged as a critical threat to autonomous driving systems. These attacks exploit the underlying neural network, allowing small -- nearly invisible -- perturbations to completely alter the behavior of such systems in potentially malicious ways. E.g., causing a traffic sign classification network to misclassify a stop sign as a speed limit sign. Prior working in hardening such systems to adversarial attacks have looked at robust training of the system or adding additional pre-processing steps to the input pipeline. Such solutions either have a hard time generalizing, require knowledge of the adversarial attacks during training, or are computationally undesirable. Instead, we propose to take insights for parameter efficient fine-tuning and use low-rank adaptation (LoRA) to train a lightweight security patch -- enabling us to dynamically patch a large preexisting vision system as new vulnerabilities are discovered. We demonstrate that our framework can patch a pre-trained model to improve classification accuracy by up to 78.01% in the presence of adversarial examples.

View on arXiv
@article{blasingame2025_2506.00661,
  title={ LoRA as a Flexible Framework for Securing Large Vision Systems },
  author={ Zander W. Blasingame and Richard E. Neddo and Chen Liu },
  journal={arXiv preprint arXiv:2506.00661},
  year={ 2025 }
}
Comments on this paper