【24h】

Posing Fair Generalization Tasks for Natural Language Inference

机译:摆出自然语言推理的公平概括任务

获取原文

摘要

Deep learning models for semantics are generally evaluated using naturalistic corpora. Adversarial methods, in which models are evaluated on new examples with known semantic properties, have begun to reveal that good performance at these naturalistic tasks can hide serious shortcomings. However, we should insist that these evaluations be fair - that the models are given data sufficient to support the requisite kinds of generalization. In this paper, we define and motivate a formal notion of fairness in this sense. We then apply these ideas to natural language inference by constructing very challenging but provably fair artificial datasets and showing that standard neural models fail to generalize in the required ways; only task-specific models that jointly compose the premise and hypothesis are able to achieve high performance, and even these models do not solve the task perfectly.
机译:语义的深度学习模型通常使用自然主义语料库进行评估。在具有已知语义属性的新示例上评估模型的对抗方法已经开始揭示,在这些自然主义任务上的良好性能可能掩盖了严重的缺陷。但是,我们应该坚持认为这些评估是公平的-为模型提供足够的数据来支持必要的一般化。在本文中,我们从这种意义上定义并激发了形式上的公平概念。然后,我们通过构建非常具有挑战性但可证明是公平的人工数据集,并证明标准神经模型无法以所需的方式进行概括,将这些思想应用于自然语言推理。只有共同构成前提和假设的特定于任务的模型才能实现高性能,甚至这些模型也不能完美地解决任务。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号