首页> 外文会议>Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies >Towards a Comprehensive Understanding and Accurate Evaluation of Societal Biases in Pre-Trained Transformers
【24h】

Towards a Comprehensive Understanding and Accurate Evaluation of Societal Biases in Pre-Trained Transformers

机译:旨在全面的理解和准确评估预训练的变压器中的社会偏见

获取原文

摘要

The ease of access to pre-trained transformers has enabled developers to leverage large-scale language models to build exciting applications for their users. While such pre-trained models offer convenient starting points for researchers and developers, there is little consideration for the societal biases captured within these model risking perpetuation of racial, gender, and other harmful biases when these models are deployed at scale. In this paper, we investigate gender and racial bias across ubiquitous pre-trained language models, including GPT-2, XLNet, BERT, RoBERTa. ALBERT and Dis-tilBERT. We evaluate bias within pre-trained transformers using three metrics: WEAT, sequence likelihood, and pronoun ranking. We conclude with an experiment demonstrating the ineffectiveness of word-embedding techniques, such as WEAT. signaling the need for more robust bias testing in transformers.
机译:易于访问预先接收的变形金刚使开发人员能够利用大规模语言模型来为其用户构建令人兴奋的应用程序。 虽然这种预先训练的模型为研究人员和开发人员提供方便的起点,但在这些模型中捕获的社会偏差很少考虑在这些模型的危险,当这些模型以比例下部署时,这些模型冒险的危险性冒险。 在本文中,我们调查了遍布培训的预先接受的语言模型的性别和种族偏见,包括GPT-2,XLNET,BERT,ROBERTA。 阿尔伯特和蒂尔伯特。 我们使用三个指标评估预先训练的变压器内的偏差:Weat,Sequiration Lotniule和代词排名。 我们通过实验结束,证明了嵌入技术的无效,例如磨损。 信号传输在变压器中需要更强大的偏置测试。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号