SOTAVerified

Assessing Intersectional Bias in Representations of Pre-Trained Image Recognition Models

2025-06-04Code Available0· sign in to hype

Valerie Krug, Sebastian Stober

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Deep Learning models have achieved remarkable success. Training them is often accelerated by building on top of pre-trained models which poses the risk of perpetuating encoded biases. Here, we investigate biases in the representations of commonly used ImageNet classifiers for facial images while considering intersections of sensitive variables age, race and gender. To assess the biases, we use linear classifier probes and visualize activations as topographic maps. We find that representations in ImageNet classifiers particularly allow differentiation between ages. Less strongly pronounced, the models appear to associate certain ethnicities and distinguish genders in middle-aged groups.

Reproductions