首页> 外文会议>Conference on empirical methods in natural language processing >Improving Cross-Lingual Word Embeddings by Meeting in the Middle
【24h】

Improving Cross-Lingual Word Embeddings by Meeting in the Middle

机译:通过在中间会面来改善交叉单词嵌入

获取原文

摘要

Cross-lingual word embeddings are becoming increasingly important in multilingual NLP. Recently, it has been shown that these embeddings can be effectively learned by aligning two disjoint monolingual vector spaces through linear transformations, using no more than a small bilingual dictionary as supervision. In this work, we propose to apply an additional transformation after the initial alignment step, which moves cross-lingual synonyms towards a middle point between them. By applying this transformation our aim is to obtain a better cross-lingual integration of the vector spaces. In addition, and perhaps surprisingly, the monolingual spaces also improve by this transformation. This is in contrast to the original alignment, which is typically learned such that the structure of the monolingual spaces is preserved. Our experiments confirm that the resulting cross-lingual embeddings outperform state-of-the-art models in both monolingual and cross-lingual evaluation tasks.
机译:跨语言嵌入式在多语言NLP中越来越重要。最近,已经示出了通过线性变换对准两个不相交的单声道向量空间,使用不超过小的双语字典作为监督来有效地学习这些嵌入。在这项工作中,我们建议在初始对齐步骤后应用额外的转换,这使跨语义同义词朝着它们之间的中间点移动。通过应用这种转变,我们的目标是获得矢量空间的更好的交叉整合。此外,令人惊讶的是,单声道空间也通过这种转变来改善。这与原始对准相反,这通常学习,使得可以保留单根空间的结构。我们的实验证实,由此产生的交叉嵌入式在单声道和交叉语言评估任务中优于最先进的模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号