【24h】

The Woman Worked as a Babysitter: On Biases in Language Generation

机译:女人当保姆:语言产生的偏见

获取原文

摘要

We present a systematic study of biases in natural language generation (NLG) by analyzing text generated from prompts that contain mentions of different demographic groups. In this work, we introduce the notion of the regard towards a demographic, use the varying levels of regard towards different demographics as a defining metric for bias in NLG, and analyze the extent to which sentiment scores are a relevant proxy metric for regard. To this end, we collect strategically-generated text from language models and manually annotate the text with both sentiment and regard scores. Additionally, we build an automatic regard classifier through transfer learning, so that we can analyze biases in unseen text. Together, these methods reveal the extent of the biased nature of language model generations. Our analysis provides a study of biases in NLG, bias metrics and correlated human judgments, and empirical evidence on the usefulness of our annotated dataset.
机译:我们通过分析从提示中包含不同人口统计群体的提示中生成的文本,对自然语言生成(NLG)中的偏见进行了系统的研究。在这项工作中,我们介绍了对人群的关注的概念,使用对不同人群的不同关注程度作为NLG偏见的定义指标,并分析了情绪得分在多大程度上是相关的关注指标。为此,我们从语言模型中收集战略性生成的文本,并用情感和关怀评分手动注释文本。此外,我们通过迁移学习构建了一个自动的关注分类器,以便我们可以分析看不见的文本中的偏见。这些方法一起揭示了语言模型世代有偏性的程度。我们的分析提供了对NLG中的偏倚,偏倚指标和相关的人类判断的研究,以及有关带注释的数据集的有用性的经验证据。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号