首页> 外文会议>International joint conference on natural language processing;Conference on empirical methods in natural language processing >Are We Modeling the Task or the Annotator? An Investigation of Annotator Bias in Natural Language Understanding Datasets
【24h】

Are We Modeling the Task or the Annotator? An Investigation of Annotator Bias in Natural Language Understanding Datasets

机译:我们是在建模任务还是在注释器上?自然语言理解数据集中注释者偏见的调查

获取原文
获取外文期刊封面目录资料

摘要

Crowdsourcing has been the prevalent paradigm for creating natural language understanding datasets in recent years. A common crowdsourcing practice is to recruit a small number of high-quality workers, and have them massively generate examples. Having only a few workers generate the majority of examples raises concerns about data diversity, especially when workers freely generate sentences. In this paper, we perform a series of experiments showing these concerns are evident in three recent NLP datasets. We show that model performance improves when training with annotator identifiers as features, and that models are able to recognize the most productive annotators. Moreover, we show that often models do not generalize well to examples from annotators that did not contribute to the training set. Our findings suggest that annotator bias should be monitored during dataset creation, and that test set annotators should be disjoint from training set annotators.
机译:近年来,众包已经成为创建自然语言理解数据集的流行范例。一种常见的众包实践是招募少量高素质的工人,并让他们大量产生榜样。仅由少数几个工人生成大多数示例会引起人们对数据多样性的担忧,尤其是当工人自由生成句子时。在本文中,我们进行了一系列实验,这些实验表明在最近的三个NLP数据集中这些担忧是显而易见的。我们表明,在将注释器标识符作为特征进行训练时,模型性能会提高,并且模型能够识别出最高效的注释器。此外,我们表明,模型通常不能很好地推广到注释者的示例中,这些注释对训练集没有帮助。我们的发现表明,在数据集创建过程中应监控注释者偏差,并且测试集注释者应与训练集注释者脱节。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号