Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings
introduction
Bias especially gender stereotypes in word embeddings:
e.g. Man - woman = programmer - homemaker
Pretrained embeddings: word2vec / 300dimensions / Google News
Quantify bias:
Compare a word vector to the vectors of a pair of gender-specific words. for example, nurse close to woman is not bias itself, because nurse close to humans, but closer than man suggest bias.
consider the distinction between gender specific words that are associated with a gender by definition (e.g. brother / sister), which close to a specfic gender is not bias, and the remaining gender neutral words (e.g. programmer / nurse).
We will use the gender specific words to learn a gender subspace ( Surprisingly, there exists a low dimensional subspace in the embedding that captures much of the gender bias.) in the embedding. Removes the bias only from the gender neutral words while respecting gender specific words.
Gender biases in English
Implicit Association Tests have uncovered gender-word biases that people do not self-report and may not even be aware of. Biases are shown in morphology as well as while there are more words referring to males, there are many more words that sexualize females than males.
Biases in algorithms
A number of online systems have been shown to exhibit various biases.Schmidt identified the bias present in word embeddings and proposed debiasing by entirely removing multiple gender dimensions. His approach is entirely remove gender from embeddings. At the same time, the difficulty of evaluating embedding quality (as compared to supervised learning) parallels the difficulty of defining bias in an embedding.
word embeddings
Embeddings form:
wϵRd
,||w||=1. Assume F-M pair
PϵRd∗Rd
, gender neutral word
NϵW
, similiarity is cosine similarity:
so similarity between embeddings is
Crowd experiments
Geometry of Gender and Bias in Word Embeddings
understand biases present in embeddings(i.e which words more close to he/she etc.) and to which extent biases agree with human notion of stereotypes.
Occupational stereotypes
Ask the crowdworkers to evaluate whether an occupation is con-sidered female-stereotypic, male-stereotypic, or neutral. Spearman r=.51(strongly correlated):
the geometric biases of embedding vectors is aligned with crowd judgment.
Analogies exhibiting stereotypes
(To Be Continued…)