首页> 外文会议>Conference of the European Chapter of the Association for Computational Linguistics >Stereotype and Skew: Quantifying Gender Bias in Pre-trained and Fine-tuned Language Models
【24h】

Stereotype and Skew: Quantifying Gender Bias in Pre-trained and Fine-tuned Language Models

机译:刻板印象和歪曲:定量预先训练和微调语言模型中的性别偏见

获取原文

摘要

This paper proposes two intuitive metrics, skew and stereotype, that quantify and analyse the gender bias present in contextual language models when tackling the WinoBias pronoun resolution task. We find evidence that gender stereotype correlates approximately negatively with gender skew in out-of-the-box models, suggesting that there is a trade-off between these two forms of bias. We investigate two methods to mitigate bias. The first approach is an online method which is effective at removing skew at the expense of stereotype. The second, inspired by previous work on ELMo, involves the fine-tuning of BERT using an augmented gender-balanced dataset. We show that this reduces both skew and stereotype relative to its unaugmented fine-tuned counterpart. However, we find that existing gender bias benchmarks do not fully probe professional bias as pronoun resolution may be obfuscated by cross-correlations from other manifestations of gender prejudice. Our code is available online.
机译:本文提出了两种直观的指标,偏斜和刻板印象,在解决Winobias代词解析任务时量化和分析语境语言模型中存在的性别偏见。 我们发现有证据表明,性别刻板印象在出箱开箱即用的模型中与性别偏差相关,这表明这两种形式的偏差之间存在权衡。 我们调查了两种缓解偏见的方法。 第一种方法是一种在线方法,它有效地以刻板印象的牺牲消除偏移。 其次受到以前的Elmo工作的推动,涉及使用增强的性别平衡数据集进行微调伯特。 我们表明,这可以减少相对于其未占用的微调对应物的偏斜和刻板印象。 但是,我们发现现有的性别偏见基准并没有完全探测专业偏见,因为代词分辨率可能会被来自其他性别偏见的其他表现形式的交叉相关性混淆。 我们的代码可在线获取。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号