首页> 外文会议>1st workshop on representation learning for NLP 2016 >Towards cross-lingual distributed representations without parallel text trained with adversarial autoencoders
【24h】

Towards cross-lingual distributed representations without parallel text trained with adversarial autoencoders

机译:在没有对抗性自动编码器训练的并行文本的情况下,实现跨语言的分布式表示

获取原文
获取原文并翻译 | 示例

摘要

Current approaches to learning vector representations of text that are compatible between different languages usually require some amount of parallel text, aligned at word, sentence or at least document level. We hypothesize however, that different natural languages share enough semantic structure that it should be possible, in principle, to learn compatible vector representations just by analyzing the monolingual distribution of words. In order to evaluate this hypothesis, we propose a scheme to map word vectors trained on a source language to vectors se-mantically compatible with word vectors trained on a target language using an adversarial autoencoder. We present preliminary qualitative results and discuss possible future developments of this technique, such as applications to cross-lingual sentence representations.
机译:当前学习不同语言之间兼容的文本向量表示的方法通常需要一定数量的平行文本,在单词,句子或至少文档级别上对齐。但是,我们假设,不同的自然语言共享足够的语义结构,原则上仅通过分析单词的单语分布,就应该有可能学习兼容的向量表示形式。为了评估该假设,我们提出了一种使用对抗性自动编码器将在源语言上训练的单词向量映射到与在目标语言上训练的单词向量语义兼容的向量的方案。我们提出了初步的定性结果,并讨论了该技术的未来发展,例如跨语言句子表示的应用。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号