首页> 外文会议>International conference on computational linguistics >Learning Word Meta-Embeddings by Autoencoding
【24h】

Learning Word Meta-Embeddings by Autoencoding

机译:通过自动编码学习单词元嵌入

获取原文

摘要

Distributed word embeddings have shown superior performances in numerous Natural Language Processing (NLP) tasks. However, their performances vary significantly across different tasks, implying that the word embeddings learnt by those methods capture complementary aspects of lexical semantics. Therefore, we believe that it is important to combine the existing word embeddings to produce more accurate and complete meta-embeddings of words. We model the meta-embedding learning problem as an autoencoding problem, where we would like to learn a meta-embedding space that can accurately reconstruct all source embeddings simultaneously. Thereby, the meta-embedding space is enforced to capture complementary information in different source embeddings via a coherent common embedding space. We propose three flavours of autoencoded meta-embeddings motivated by different requirements that must be satisfied by a meta-embedding. Our experimental results on a series of benchmark evaluations show that the proposed auloencoded meta-embeddings outperform the existing slate-of-the-art meta-embeddings in multiple tasks.
机译:在许多自然语言处理(NLP)任务中,分布式词嵌入已显示出卓越的性能。但是,它们在不同任务之间的性能差异很大,这意味着通过这些方法学习的词嵌入可以捕获词汇语义的互补方面。因此,我们认为将现有的单词嵌入组合在一起以产生更准确和完整的单词元嵌入非常重要。我们将元嵌入学习问题建模为自动编码问题,在此我们希望学习一个元嵌入空间,该空间可以同时准确地重建所有源嵌入。从而,元嵌入空间被强制通过相干的公共嵌入空间来捕获不同源嵌入中的补充信息。我们提出了三种风味的自动编码元嵌入,这些元嵌入是由元嵌入必须满足的不同要求所激发的。我们在一系列基准评估上的实验结果表明,在多个任务中,拟议的全编码元嵌入优于现有的最新的元嵌入。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号