【24h】

Learning Dense Representations of Phrases at Scale

机译:在规模上学习密集的短语表示

获取原文

摘要

Open-domain question answering can be reformulated as a phrase retrieval problem, without the need for processing documents on-demand during inference (Seo et al., 2019). However, current phrase retrieval models heavily depend on sparse representations and still underper-fonn retriever-reader approaches. In this work, we show for the first time that we can learn dense representations of phrases alone that achieve much stronger performance in open-domain QA. We present an effective method to learn phrase representations from the supervision of reading comprehension tasks, coupled with novel negative sampling methods. We also propose a query-side line-tuning strategy, which can support transfer learning and reduce the discrepancy between training and inference. On five popular open-domain QA datasets, our model DensePhrases improves over previous phrase retrieval models by 15%- 25% absolute accuracy and matches the performance of state-of-the-art retriever-reader models. Our model is easy to parallelize due to pure dense representations and processes more than 10 questions per second on CPUs. Finally, we directly use our pre-indexed dense phrase representations for two slot filling tasks, showing the promise of utilizing DensePhrases as a dense knowledge base for downstream tasks.
机译:开放式域问题回答可以重新重新重新重新重新检索问题,无需在推理期间按需处理文档(SEO等,2019)。然而,当前短语检索模型严重依赖于稀疏表示,并且仍未欠下较低的鼠尾读者方法。在这项工作中,我们首次展示我们可以在开放式QA中学习在开放域QA中实现更强烈的表现的密集表现。我们提出了一种学习短语表示从阅读理解任务的监督的陈述的有效方法,与新颖的负抽样方法相结合。我们还提出了一个查询侧的线路调整策略,可以支持转移学习并降低培训和推理之间的差异。在五个流行的开放式QA数据集上,我们的模型DendePhrases通过前一词的检索模型提高了15% - 25%的绝对精度,并符合最先进的检索器读者型号的性能。由于纯密集的表示,我们的模型很容易并行化,并在CPU上每秒处理超过10个问题。最后,我们直接使用我们的预索引密度的密度短语表示两个插槽填充任务,显示使用DendePhrase作为下游任务的密集知识库的承诺。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号