SOTAVerified

Measuring Gender Bias in Word Embeddings across Domains and Discovering New Gender Bias Word Categories

2019-08-01WS 2019Code Available0· sign in to hype

Kaytlin Chaloner, Alfredo Maldonado

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Prior work has shown that word embeddings capture human stereotypes, including gender bias. However, there is a lack of studies testing the presence of specific gender bias categories in word embeddings across diverse domains. This paper aims to fill this gap by applying the WEAT bias detection method to four sets of word embeddings trained on corpora from four different domains: news, social networking, biomedical and a gender-balanced corpus extracted from Wikipedia (GAP). We find that some domains are definitely more prone to gender bias than others, and that the categories of gender bias present also vary for each set of word embeddings. We detect some gender bias in GAP. We also propose a simple but novel method for discovering new bias categories by clustering word embeddings. We validate this method through WEAT's hypothesis testing mechanism and find it useful for expanding the relatively small set of well-known gender bias word categories commonly used in the literature.

Tasks

Reproductions