首页> 外文会议>International Symposium on Telecommunications >Persian Language Modeling Using Recurrent Neural Networks
【24h】

Persian Language Modeling Using Recurrent Neural Networks

机译:波斯语言建模使用经常性神经网络

获取原文

摘要

In this paper, recurrent neural networks are applied to language modeling of Persian, using word embedding as word representation. To this aim, unidirectional and bidirectional Long Short-Term Memory (LSTM) networks are used, and the perplexity of Persian language models on a 100-million-word data set is evaluated. The effect of various parameters, including number of hidden layers and size of LSTM units, on the performance of the networks in reducing the perplexity of the models are investigated. Among different LSTM language models, the best perplexity, which is equal to 59.05, is achieved from a 2-layer bidirectional LSTM model. Comparing this value with the perplexity of the classical Trigram model, which is equal to 138, an improvement in the modeling is noticeable, which is due to the ability of neural networks to make a higher generalization in comparison with the well-known N-gram model.
机译:在本文中,经常性神经网络应用于波斯语建模,使用Word Embedding作为Word表示。为此目的,使用单向和双向长期内存(LSTM)网络,评估在100亿字数据集上的波斯语模型的困惑。研究了各种参数,包括隐藏层数和LSTM单元的尺寸的效果,用于降低模型的困惑的网络的性能。在不同的LSTM语言模型中,从2层双向LSTM模型实现了等于59.05的最佳困惑。将该值与等于138等于138的经典三元模型的困惑,这是显着的,这是由于神经网络与众所周知的n-gram相比具有更高概括的能力,这是由于神经网络的能力模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号