【24h】

BoolQ: Exploring the Surprising Difficulty of Natural Yes/No Questions

机译:BoolQ:探索自然是/否问题的惊人难度

获取原文

摘要

In this paper we study yeso questions that are naturally occurring - meaning that they are generated in unprompted and unconstrained settings. We build a reading comprehension dataset, BoolQ, of such questions, and show that they are unexpectedly challenging. They often query for complex, non-factoid information, and require difficult entailment-like inference to solve. We also explore the effectiveness of a range of transfer learning baselines. We find that transferring from entailment data is more effective than transferring from paraphrase or extractive QA data, and that it, surprisingly, continues to be very beneficial even when starting from massive pre-trained language models such as BKRT. Our best method trains BERT on MultiNLI and then re-trains it on our train set. It achieves 80.4% accuracy compared to 90% accuracy of human anno-tators (and 62% majority-baseline), leaving a significant gap for future work.
机译:在本文中,我们研究自然会发生的是/否问题,这意味着它们是在无提示且不受限制的情况下生成的。我们建立了此类问题的阅读理解数据集BoolQ,并表明它们出乎意料地具有挑战性。他们经常查询复杂的,非事实的信息,并且需要困难的类似蕴含的推理来解决。我们还探讨了一系列迁移学习基准的有效性。我们发现从包含数据传输比从释义或提取质量检查数据传输更有效,而且令人惊讶的是,即使从大规模的预训练语言模型(例如BKRT)开始,它仍然非常有益。我们最好的方法是在MultiNLI上训练BERT,然后在我们的火车上对其进行重新训练。与人类注释者的90%的准确性(以及62%的多数基线)相比,它的准确性达到80.4%,为以后的工作留下了很大的空白。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号