Notes_Man2Programmer@Woman2Homemaker

Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings

introduction
Bias especially gender stereotypes in word embeddings:

e.g. Man - woman = programmer - homemaker

Pretrained embeddings: word2vec / 300dimensions / Google News

Quantify bias:

Compare a word vector to the vectors of a pair of gender-specific words. for example, nurse close to woman is not bias itself, because nurse close to humans, but closer than man suggest bias.

consider the distinction between gender specific words that are associated with a gender by definition (e.g. brother / sister), which close to a specfic gender is not bias, and the remaining gender neutral words (e.g. programmer / nurse).

We will use the gender specific words to learn a gender subspace ( Surprisingly, there exists a low dimensional subspace in the embedding that captures much of the gender bias.) in the embedding. Removes the bias only from the gender neutral words while respecting gender specific words.

Gender biases in English

Implicit Association Tests have uncovered gender-word biases that people do not self-report and may not even be aware of. Biases are shown in morphology as well as while there are more words referring to males, there are many more words that sexualize females than males.

Biases in algorithms

A number of online systems have been shown to exhibit various biases.Schmidt identified the bias present in word embeddings and proposed debiasing by entirely removing multiple gender dimensions. His approach is entirely remove gender from embeddings. At the same time, the difficulty of evaluating embedding quality (as compared to supervised learning) parallels the difficulty of defining bias in an embedding.

word embeddings

Embeddings form: wϵRd ,||w||=1. Assume F-M pair PϵRdRd , gender neutral word NϵW , similiarity is cosine similarity:

cos(u,v)=uv|u||v|

so similarity between embeddings is
cos(w1,w2)=w1w2(2)

Crowd experiments
Geometry of Gender and Bias in Word Embeddings

understand biases present in embeddings(i.e which words more close to he/she etc.) and to which extent biases agree with human notion of stereotypes.

Occupational stereotypes

Ask the crowdworkers to evaluate whether an occupation is con-sidered female-stereotypic, male-stereotypic, or neutral. Spearman r=.51(strongly correlated):

the geometric biases of embedding vectors is aligned with crowd judgment.

Analogies exhibiting stereotypes

(To Be Continued…)

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值