42
0

Spurious Privacy Leakage in Neural Networks

Main:9 Pages
10 Figures
Bibliography:6 Pages
10 Tables
Appendix:7 Pages
Abstract

Neural networks are vulnerable to privacy attacks aimed at stealing sensitive data. The risks can be amplified in a real-world scenario, particularly when models are trained on limited and biased data. In this work, we investigate the impact of spurious correlation bias on privacy vulnerability. We introduce \emph{spurious privacy leakage}, a phenomenon where spurious groups are significantly more vulnerable to privacy attacks than non-spurious groups. We further show that group privacy disparity increases in tasks with simpler objectives (e.g. fewer classes) due to the persistence of spurious features. Surprisingly, we find that reducing spurious correlation using spurious robust methods does not mitigate spurious privacy leakage. This leads us to introduce a perspective on privacy disparity based on memorization, where mitigating spurious correlation does not mitigate the memorization of spurious data, and therefore, neither the privacy level. Lastly, we compare the privacy of different model architectures trained with spurious data, demonstrating that, contrary to prior works, architectural choice can affect privacy outcomes.

View on arXiv
@article{zhang2025_2505.20095,
  title={ Spurious Privacy Leakage in Neural Networks },
  author={ Chenxiang Zhang and Jun Pang and Sjouke Mauw },
  journal={arXiv preprint arXiv:2505.20095},
  year={ 2025 }
}
Comments on this paper