首页> 外文会议>Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies >RECONSIDER: Improved Re-Ranking using Span-Focused Cross-Attention for Open Domain Question Answering
【24h】

RECONSIDER: Improved Re-Ranking using Span-Focused Cross-Attention for Open Domain Question Answering

机译:重新考虑:使用跨度侧面关注的开放域问题回答改进重新排名

获取原文

摘要

State-of-the-art Machine Reading Comprehension (MRC) models for Open-domain Question Answering (QA) are typically trained for span selection using distantly supervised positive examples and heuristically retrieved negative examples. This training scheme possibly explains empirical observations that these models achieve a high recall amongst their top few predictions, but a low overall accuracy, motivating the need for answer re-ranking. We develop a successful re-ranking approach (RECONSIDER) for span-extraction tasks that improves upon the performance of MRC models, even beyond large-scale pre-training. Re-Consider is trained on positive and negative examples extracted from high confidence MRC model predictions, and uses in-passage span annotations to perform span-focused re-ranking over a smaller candidate set. As a result, RECONSIDER learns to eliminate close false positives, achieving a new extractive state of the art on four QA tasks, with 45.5% Exact Match accuracy on Natural Questions with real user questions, and 61.7% on TriviaQA. We will release all related data, models, and code.
机译:最先进的机器阅读理解(MRC)用于开放式域问题应答(QA)的模型通常使用远端监督的正示例和启发式检索的负例进行跨度选择培训。该培训方案可能解释了这些模型在其顶层预测中实现了高召回的实证观察,但总体准确性低,激励了回复重新排名的需求。我们开发成功的重新排名方法(重新考虑),用于跨度提取任务,即使超越大规模预训练,也可以提高MRC模型的性能。重新考虑在高置信频率MRC模型预测中提取的正面和否定示例培训,并使用内部跨度注释来执行以较小的候选集进行跨度重新排序。因此,重新考虑学会消除近距离误报,在四个QA任务中实现了最新的现有技术,对自然问题有45.5%的精确度,具有真实的用户问题,在TriviaQA上有61.7%。我们将释放所有相关数据,模型和代码。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号