首页> 外文会议>Conference on Neural Information Processing Systems >Assessing Social and Intersectional Biases in Contextualized Word Representations
【24h】

Assessing Social and Intersectional Biases in Contextualized Word Representations

机译:评估语境化词表示中的社会和交叉偏见

获取原文
获取外文期刊封面目录资料

摘要

Social bias in machine learning has drawn significant attention, with work ranging from demonstrations of bias in a multitude of applications, curating definitions of fairness for different contexts, to developing algorithms to mitigate bias. In natural language processing, gender bias has been shown to exist in context-free word embeddings. Recently, contextual word representations have outperformed word embeddings in several downstream NLP tasks. These word representations are conditioned on their context within a sentence, and can also be used to encode the entire sentence. In this paper, we analyze the extent to which state-of-the-art models for contextual word representations, such as BERT and GPT-2, encode biases with respect to gender, race, and intersectional identities. Towards this, we propose assessing bias at the contextual word level. This novel approach captures the contextual effects of bias missing in context-free word embeddings, yet avoids confounding effects that underestimate bias at the sentence encoding level. We demonstrate evidence of bias at the corpus level, find varying evidence of bias in embedding association tests, show in particular that racial bias is strongly encoded in contextual word models, and observe that bias effects for intersectional minorities are exacerbated beyond their constituent minority identities. Further, evaluating bias effects at the contextual word level captures biases that are not captured at the sentence level, confirming the need for our novel approach.
机译:机器学习中的社会偏见引起了重大关注,工作范围从偏见的偏见示范中的偏见,对不同背景的公平定义,发展算法以减轻偏差。在自然语言处理中,在无背景中的单词嵌入中显示了性别偏见。最近,上下文字表示在几个下游NLP任务中具有表现优于嵌入式。这些字表示在句子中的上下文上有调节,也可用于编码整个句子。在本文中,我们分析了诸如BERT和GPT-2的最新模型,例如BERT和GPT-2,编码关于性别,种族和交叉标识的偏差的最新模型。为此,我们建议在语境字级别评估偏见。这种新颖的方法捕获了无背景中缺失的偏差的上下文影响,但避免了在句子编码级别低估的混淆效果。我们展示了语料库级别的偏见的证据,发现嵌入关联测试中的偏差的不同证据,特别是在上下文中的语境模型中强烈编码种族偏差,并观察到交叉少数群体的偏差效应超出其组成少数群体。此外,在上下文中的偏差效应评估偏差效应捕获在句子级别捕获的偏差,确认需要我们的新方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号