【24h】

Learning to Rank for Plausible Plausibility

机译:学习为合理的合理性等级

获取原文
获取外文期刊封面目录资料

摘要

Researchers illustrate improvements in contextual encoding strategies via resultant performance on a battery of shared Natural Language Understanding (NLU) tasks. Many of these tasks are of a categorical prediction variety: given a conditioning context (e.g., an NLI premise), provide a label based on an associated prompt (e.g., an NLI hypothesis). The categorical nature of these tasks has led to common use of a cross entropy log-loss objective during training. We suggest this loss is intuitively wrong when applied to plausibility tasks, where the prompt by design is neither categorically entailed nor contradictory given the context. Log-loss naturally drives models to assign scores near 0.0 or 1.0, in contrast to our proposed use of a margin-based loss. Following a discussion of our intuition, we describe a confirmation study based on an extreme, synthetically curated task derived from MultiNLI. We find that a margin-based loss leads to a more plausible model of plausibility. Finally, we illustrate improvements on the Choice Of Plausible Alternative (COPA) task through this change in loss.
机译:研究人员说明了通过对共享自然语言理解(NLU)任务的电池产生的性能进行上下文编码策略的改进。许多这些任务是一个分类预测品种:给定调节上下文(例如,NLI之前),提供基于相关提示的标签(例如,NLI假设)。这些任务的分类性质导致培训期间常用于交叉熵数损失目标。我们建议在适用于合理性任务时,这种损失是直观的错误,其中设计的提示既不鉴于上下文,既不明显也不矛盾。与我们建议使用基于利润的损失,日志丢失自然驱动模型以分配近0.0或1.0附近的分数。在讨论我们的直觉之后,我们描述了一种基于来自Multinli的极端综合策划任务的确认研究。我们发现基于保证金的损失导致更合理的合理性模型。最后,我们通过这种损失变化说明了对合理替代(COPA)任务的选择的改进。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号