SOTAVerified

Detecting Independent Pronoun Bias with Partially-Synthetic Data Generation

2020-11-01EMNLP 2020Unverified0· sign in to hype

Robert Munro, Alex (Carmen) Morrison

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

We report that state-of-the-art parsers consistently failed to identify ``hers'' and ``theirs'' as pronouns but identified the masculine equivalent ``his''. We find that the same biases exist in recent language models like BERT. While some of the bias comes from known sources, like training data with gender imbalances, we find that the bias is \_amplified\_ in the language models and that linguistic differences between English pronouns that are not inherently biased can become biases in some machine learning models. We introduce a new technique for measuring bias in models, using Bayesian approximations to generate partially-synthetic data from the model itself.

Tasks

Reproductions