首页> 外文会议>Conference on empirical methods in natural language processing >Improving Cross-Lingual Word Embeddings by Meeting in the Middle
【24h】

Improving Cross-Lingual Word Embeddings by Meeting in the Middle

机译:通过中间会议改善跨语言单词嵌入

获取原文

摘要

Cross-lingual word embeddings are becoming increasingly important in multilingual NLP. Recently, it has been shown that these embeddings can be effectively learned by aligning two disjoint monolingual vector spaces through linear transformations, using no more than a small bilingual dictionary as supervision. In this work, we propose to apply an additional transformation after the initial alignment step, which moves cross-lingual synonyms towards a middle point between them. By applying this transformation our aim is to obtain a better cross-lingual integration of the vector spaces. In addition, and perhaps surprisingly, the monolingual spaces also improve by this transformation. This is in contrast to the original alignment, which is typically learned such that the structure of the monolingual spaces is preserved. Our experiments confirm that the resulting cross-lingual embeddings outperform state-of-the-art models in both monolingual and cross-lingual evaluation tasks.
机译:跨语言单词嵌入在多语言NLP中变得越来越重要。最近,已经证明,仅使用一个小的双语词典作为监督,通过线性变换排列两个不相交的单语向量空间,就可以有效地学习这些嵌入。在这项工作中,我们建议在初始对齐步骤之后应用一个附加的转换,该转换将跨语言的同义词移向它们之间的中间点。通过应用此变换,我们的目标是获得向量空间的更好的跨语言集成。另外,也许令人惊讶的是,这种转换也使单语空间得到了改善。这与原始对准相反,原始对准通常被学习使得保留单语空间的结构。我们的实验证实,所产生的跨语言嵌入在单语言和跨语言评估任务中均优于最新模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号