首页> 外文会议>Conference on computational natural language learning >Better Word Representations with Recursive Neural Networks for Morphology
【24h】

Better Word Representations with Recursive Neural Networks for Morphology

机译:递归神经网络用于形态学的更好的单词表示

获取原文

摘要

Vector-space word representations have been very successful in recent years at improving performance across a variety of NLP tasks. However, common to most existing work, words are regarded as independent entities without any explicit relationship among morphologically related words being modeled. As a result, rare and complex words are often poorly estimated, and all unknown words are represented in a rather crude way using only one or a few vectors. This paper addresses this shortcoming by proposing a novel model that is capable of building representations for morphologically complex words from their morphemes. We combine recursive neural networks (RNNs), where each morpheme is a basic unit, with neural language models (NLMs) to consider contextual information in learning morphologically-aware word representations. Our learned models outperform existing word representations by a good margin on word similarity tasks across many datasets, including a new dataset we introduce focused on rare words to complement existing ones in an interesting way.
机译:近年来,向量空间字表示在提高各种NLP任务的性能方面已经非常成功。但是,对于大多数现有工作而言,单词被视为独立实体,而在建模上与词法相关的单词之间没有任何明确的关系。结果,稀疏和复杂的单词常常被错误地估计,并且所有未知单词仅使用一个或几个向量就以相当粗略的方式表示。本文通过提出一种新颖的模型来解决这一缺点,该模型能够从词素中构造出形态复杂的单词。我们将递归神经网络(RNN)(其中每个词素是一个基本单元)与神经语言模型(NLM)相结合,以在学习形态感知词表示形式时考虑上下文信息。我们的学习模型在许多数据集上的单词相似性任务上都比现有单词表示好得多,其中包括我们引入的新数据集,重点关注稀有单词,以一种有趣的方式对现有单词进行补充。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号