首页> 外文会议>International conference on recent advances in natural language processing >Meta-Embedding Sentence Representation for Textual Similarity
【24h】

Meta-Embedding Sentence Representation for Textual Similarity

机译:文本相似性的元嵌入句子表示

获取原文

摘要

Word embedding models are now widely used in most NLP applications. Despite their effectiveness, there is no clear evidence about the choice of the most appropriate model. It often depends on the nature of the task and on the quality and size of the used data sets. This remains true for bottom-up sentence embedding models. However, no straightforward investigation has been conducted so far. In this paper, we propose a systematic study of the impact of the main word embedding models on sentence representation. By contrasting in-domain and pre-trained embedding models, we show under which conditions they can be jointly used for bottom-up sentence embeddings. Finally, we propose the first bottom-up meta-embedding representation at the sentence level for textual similarity. Significant improvements are observed in several tasks including question-to-question similarity, paraphrasing and next utterance ranking.
机译:现在,单词嵌入模型已在大多数NLP应用程序中广泛使用。尽管它们有效,但是没有关于选择最合适模型的明确证据。它通常取决于任务的性质以及所使用数据集的质量和大小。对于自下而上的句子嵌入模型,情况仍然如此。但是,到目前为止,尚未进行直接调查。在本文中,我们提出了对主要词嵌入模型对句子表示的影响的系统研究。通过对比域内和预训练的嵌入模型,我们显示了在什么条件下它们可以联合用于自下而上的句子嵌入。最后,出于文本相似性,我们在句子级别提出了第一个自下而上的元嵌入表示形式。在几个任务中观察到了显着的改进,包括问题之间的相似性,措辞和下一个话语排名。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号