v1v2 (latest)
Assessing Intersectional Bias in Representations of Pre-Trained Image Recognition Models

Main:17 Pages
11 Figures
Bibliography:3 Pages
5 Tables
Abstract
Deep Learning models have achieved remarkable success. Training them is often accelerated by building on top of pre-trained models which poses the risk of perpetuating encoded biases. Here, we investigate biases in the representations of commonly used ImageNet classifiers for facial images while considering intersections of sensitive variables age, race and gender. To assess the biases, we use linear classifier probes and visualize activations as topographic maps. We find that representations in ImageNet classifiers particularly allow differentiation between ages. Less strongly pronounced, the models appear to associate certain ethnicities and distinguish genders in middle-aged groups.
View on arXiv@article{krug2025_2506.03664, title={ Assessing Intersectional Bias in Representations of Pre-Trained Image Recognition Models }, author={ Valerie Krug and Sebastian Stober }, journal={arXiv preprint arXiv:2506.03664}, year={ 2025 } }
Comments on this paper