首页> 外文会议>First workshop on subword and character level models in NLP 2017 >Character and Subword-Based word Representation for Neural Language Modeling prediction
【24h】

Character and Subword-Based word Representation for Neural Language Modeling prediction

机译:基于字符和子词的词表示在神经语言建模预测中的应用

获取原文
获取原文并翻译 | 示例

摘要

Most of neural language models use different kinds of embeddings for word prediction. While word embeddings can be associated to each word in the vocabulary or derived from characters as well as factored morphological decomposition, these word representations are mainly used to parametrize the input, i.e. the context of prediction. This work investigates the effect of using subword units (character and factored morphological decomposition) to build output representations for neural language modeling. We present a case study on Czech, a morphologically-rich language, experimenting with different input and output representations. When working with the full training vocabulary, despite unstable training, our experiments show that augmenting the output word representations with character-based embeddings can significantly improve the performance of the model. Moreover, reducing the size of the output look-up table, to let the character-based embeddings represent rare words, brings further improvement.
机译:大多数神经语言模型使用不同类型的嵌入进行单词预测。虽然词嵌入可以与词汇表中的每个词相关联,或者可以从字符中派生出来,以及可以进行因式分解,但这些词表示法主要用于对输入进行参数化,即预测的上下文。这项工作研究了使用子词单元(字符和分解的词法分解)来构建用于神经语言建模的输出表示的效果。我们提供了一个捷克语的案例研究,捷克语是一种形态丰富的语言,并尝试了不同的输入和输出表示形式。当使用完整的训练词汇时,尽管训练不稳定,我们的实验表明,使用基于字符的嵌入来增强输出单词表示形式可以显着提高模型的性能。此外,减小输出查找表的大小,以使基于字符的嵌入表示稀有单词,带来了进一步的改进。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号