首页> 外文会议>Annual meeting of the Association for Computational Linguistics >Compressing Neural Language Models by Sparse Word Representations
【24h】

Compressing Neural Language Models by Sparse Word Representations

机译:通过稀疏词表示压缩神经语言模型

获取原文
获取外文期刊封面目录资料

摘要

Neural networks are among the state-of-the-art techniques for language modeling. Existing neural language models typically map discrete words to distributed, dense vector representations. After information processing of the preceding context words by hidden layers, an output layer estimates the probability of the next word. Such approaches are time- and memory-intensive because of the large numbers of parameters for word embeddings and the output layer. In this paper, we propose to compress neural language models by sparse word representations. In the experiments, the number of parameters in our model increases very slowly with the growth of the vocabulary size, which is almost imperceptible. Moreover, our approach not only reduces the parameter space to a large extent, but also improves the performance in terms of the perplexity measure.
机译:神经网络是语言建模的最新技术之一。现有的神经语言模型通常将离散的单词映射到分布式的密集矢量表示。在通过隐藏层对前面的上下文词进行信息处理之后,输出层估计下一个词的概率。由于用于词嵌入和输出层的大量参数,所以这种方法是时间和存储器密集型的。在本文中,我们建议通过稀疏词表示来压缩神经语言模型。在实验中,我们模型中参数的数量随着词汇量的增加而非常缓慢地增加,这几乎是不可察觉的。此外,我们的方法不仅在很大程度上减少了参数空间,而且在困惑度方面也提高了性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号