【24h】

Obtaining Better Word Representations via Language Transfer

机译:通过语言转移获得更好的单词表示

获取原文

摘要

Vector space word representations have gained big success recently at improving performance across various NLP tasks. However, existing word embeddings learning methods only utilize homo-lingual corpus. Inspired by transfer learning, we propose a novel language transfer method to obtain word embeddings via language transfer. Under this method, in order to obtain word embeddings of one language (target language), we train models on corpus of another different language (source language) instead. And then we use the obtained source language word embeddings to represent target language word embeddings. We evaluate the word embeddings obtained by the proposed method on word similarity tasks across several benchmark datasets. And the results show that our method is surprisingly effective, outperforming competitive baselines by a large margin. Another benefit of our method is that the process of collecting new corpus might be skipped.
机译:传染媒介空间字表示最近在提高各种NLP任务中提高了表现的大量成功。 但是,现有的Word Embeddings学习方法仅利用同性恋语料库。 通过转移学习的启发,我们提出了一种新颖的语言转移方法,通过语言转移获取Word Embedings。 在此方法下,为了获得一种语言(目标语言)的单词嵌入,我们培训另一种不同语言(源语言)的语料库上的模型。 然后我们使用所获得的源语言Word Embeddings表示目标语言单词嵌入品。 我们评估通过在多个基准数据集中的单词相似性任务上获得的嵌入单词。 结果表明,我们的方法令人惊讶地有效,优于竞争力的基线。 我们的方法的另一个好处是,可能会跳过收集新语料库的过程。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号