首页> 外文会议>International Conference on Text, Speech and Dialogue >Continuous Distributed Representations of Words as Input of LSTM Network Language Model
【24h】

Continuous Distributed Representations of Words as Input of LSTM Network Language Model

机译:单词的连续分布式表示作为LSTM网络语言模型的输入

获取原文

摘要

The continuous skip-gram model is an efficient algorithm for learning quality distributed vector representations that are able to capture a large number of syntactic and semantic word relationships. Artificial neural networks have become the state-of-the-art in the task of language modelling whereas Long-Short Term Memory (LSTM) networks seem to be efficient training algorithm. In this paper, we carry out experiments with a combination of these powerful models: the continuous distributed representations of words are trained with skip-gram method on a big corpora and are used as the input of LSTM language model instead of traditional 1-of-N coding. The possibilities of this approach are shown in experiments on perplexity with Wikipedia and Penn Treebank corpus.
机译:连续的Skip-Gram模型是一种高效的学习质量分布式矢量表示的算法,可以捕获大量的句法和语义词关系。在语言建模的任务中,人工神经网络已成为最先进的,而长短短期内存(LSTM)网络似乎是有效的训练算法。在本文中,我们对这些强大模型的组合进行了实验:单词的连续分布式表示在大公司的跳过革克方法中培训,用作LSTM语言模型的输入而不是传统的1-of-编码。这种方法的可能性在与维基百科和Penn TreeBank语料库困惑的实验中显示。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号