首页> 外文OA文献 >Bilingual word embeddings from non-parallel document-aligned data applied to bilingual lexicon induction
【2h】

Bilingual word embeddings from non-parallel document-aligned data applied to bilingual lexicon induction

机译:来自非平行文档对齐数据的双语词嵌入应用于双语词典归纳

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

We propose a simple yet effective approach to learning bilingual word embeddings (BWEs) from non-parallel document-aligned data (based on the omnipresent skip-gram model), and its application to bilingual lexicon induction (BLI). We demonstrate the utility of the induced BWEs in the BLI task by reporting on benchmarking BLI datasets for three language pairs: (1) We show that our BWE-based BLI models significantly outperform the MuPTM-based and context-counting models in this setting, and obtain the best reported BLI results for all three tested language pairs; (2) We also show that our BWE-based BLI models outperform other BLI models based on recently proposed BWEs that require parallel data for bilingual training.
机译:我们提出了一种简单而有效的方法,用于从非并行文档对齐数据(基于无所不在的跳过语法模型)学习双语词嵌入(BWE),并将其应用于双语词典归纳(BLI)。通过报告三种语言对的基准BLI数据集,我们证明了诱导BWE在BLI任务中的效用:(1)我们证明了在这种情况下,基于BWE的BLI模型明显优于基于MuPTM和上下文计数的模型,并获得所有三种测试语言对报告的最佳BLI结果; (2)我们还表明,基于BWE的BLI模型优于基于最近提出的BWE的其他BLI模型,这些BWE需要并行数据进行双语培训。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号