首页> 外文会议>Annual meeting of the Association for Computational Linguistics >Unsupervised Cross-Lingual Representation Learning
【24h】

Unsupervised Cross-Lingual Representation Learning

机译:无监督的跨语言表示学习

获取原文

摘要

Cross-lingual word representations offer an elegant and language-pair independent way to represent content across different languages. They enable us to reason about word meaning in multilingual contexts and serve as an integral source of knowledge for multilingual applications such as machine translation (Artetxe et al., 2018d; Qi et al., 2018; Lample et al., 2018b) or multilingual search and question answering (Vulic and Moens, 2015). In addition, they are a key facilitator of cross-lingual transfer and joint multilingual training, offering support to NLP applications in a large spectrum of languages (Søgaard et al., 2015; Ammar et al., 2016a). While NLP is increasingly more embedded into a variety of products related to, e.g., translation, conversational or search tasks, resources such as annotated training data are still lacking or insufficient to induce satisfying models for many resource-poor languages. There arc often no trained linguistic annotators for these languages, and markets may be too small or premature to invest in such training. This is a major challenge, but cross-lingual modelling and transfer can help by exploiting observable correlations between major languages and low-resource languages.
机译:跨语言单词表示提供了一种优雅且独立于语言对的方式来表示不同语言的内容。它们使我们能够在多语言环境中推理单词的含义,并成为机器翻译(Artetxe等人,2018d; Qi等人,2018; Lample等人,2018b)或多语言应用程序等多语言应用程序的不可或缺的知识来源搜索和问题解答(Vulic和Moens,2015年)。此外,它们是跨语言迁移和联合多语言培训的主要促进者,为多种语言的NLP应用程序提供了支持(Søgaard等,2015; Ammar等,2016a)。尽管NLP越来越多地嵌入到与翻译,会话或搜索任务等相关的各种产品中,但是诸如注释的训练数据之类的资源仍然缺乏或不足以为许多资源贫乏的语言带来令人满意的模型。对于这些语言,通常没有经过培训的语言注释者,并且市场规模可能太小或为时过早,无法在此类培训上进行投资。这是一个重大挑战,但是跨语言的建模和传递可以通过利用主要语言和资源匮乏的语言之间的可观察到的关联来提供帮助。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号