【24h】

A Question-Focused Multi-Factor Attention Network for Question Answering

机译:一个有问题的多因素注意网络,用于问答

获取原文

摘要

Neural network models recently proposed for question answering (QA) primarily focus on capturing the passage-question relation. However, they have minimal capability to link relevant facts distributed across multiple sentences which is crucial in achieving deeper understanding, such as performing multi-sentence reasoning, co-reference resolution, etc. They also do not explicitly focus on the question and answer type which often plays a critical role in QA. In this paper, we propose a novel end-to-end question-focused multi-factor attention network for answer extraction. Multi-factor attentive encoding using tensor-based transformation aggregates meaningful facts even when they are located in multiple sentences. To implicitly infer the answer type, we also propose a max-attentional question aggregation mechanism to encode a question vector based on the important words in a question. During prediction, we incorporate sequence-level encoding of the first wh-word and its immediately following word as an additional source of question type information. Our proposed model achieves significant improvements over the best prior state-of-the-art results on three large-scale challenging QA datasets, namely NewsQA, TriviaQA, and SearchQA.
机译:最近提出的神经网络模型提出了问题应答(QA)主要集中在捕获段询问题。然而,它们具有最小的能力,可以将分布的多句话中的相关事实联系在实现更深层次的理解中,例如执行多句子推理,共同参考分辨率等。它们也没有明确关注问题和答案类型经常在QA中发挥关键作用。在本文中,我们提出了一种用于回答提取的新颖的端到端问题的多因素注意网络。即使它们位于多个句子中,使用基于张量的转换的多因素细分编码也是有意义的事实。为了隐含地推断出答案类型,我们还提出了一种最大注意力问题,以基于问题中的重要单词编码问题载体。在预测期间,我们将第一个wh-word的序列级编码纳入了一个单词作为额外的问题类型信息来源。我们所提出的模型在三个大型挑战性QA数据集,即Newsqa,TriviaQA和SearchQA上实现了最先进的最先进结果的显着改进。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号