首页> 外文期刊>Machine translation >The representational geometry of word meanings acquired by neural machine translation models
【24h】

The representational geometry of word meanings acquired by neural machine translation models

机译:神经机器翻译模型获得的词义表示几何

获取原文
获取原文并翻译 | 示例
           

摘要

This work is the first comprehensive analysis of the properties of word embeddings learned by neural machine translation (NMT) models trained on bilingual texts. We show the word representations of NMT models outperform those learned from monolingual text by established algorithms such as Skipgram and CBOW on tasks that require knowledge of semantic similarity and/or lexical-syntactic role. These effects hold when translating from English to French and English to German, and we argue that the desirable properties of NMT word embeddings should emerge largely independently of the source and target languages. Further, we apply a recently-proposed heuristic method for training NMT models with very large vocabularies, and show that this vocabulary expansion method results in minimal degradation of embedding quality. This allows us to make a large vocabulary of NMT embeddings available for future research and applications. Overall, our analyses indicate that NMT embeddings should be used in applications that require word concepts to be organised according to similarity and/or lexical function, while monolingual embeddings are better suited to modelling (nonspecific) inter-word relatedness.
机译:这项工作是对通过双语文本训练的神经机器翻译(NMT)模型学习的词嵌入属性的首次全面分析。我们显示,在需要知识相似性和/或词汇句法作用知识的任务上,NMT模型的单词表示优于通过已建立的算法(例如Skipgram和CBOW)从单语文本中学习的单词表示。从英语到法语和英语到德语翻译时,这些效果仍然存在。我们认为NMT词嵌入的理想属性应在很大程度上独立于源语言和目标语言而出现。此外,我们应用了最近提出的启发式方法来训练具有非常大的词汇量的NMT模型,并证明了这种词汇量扩展方法可以最大程度地降低嵌入质量。这使我们能够为将来的研究和应用提供大量的NMT嵌入词汇。总体而言,我们的分析表明,在需要根据相似性和/或词汇功能来组织词概念的应用中应使用NMT嵌入,而单语言嵌入更适合于建模(非特定的)词间相关性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号