首页> 外文期刊>International journal of speech technology >Enhancing accuracy of long contextual dependencies for Punjabi speech recognition system using deep LSTM
【24h】

Enhancing accuracy of long contextual dependencies for Punjabi speech recognition system using deep LSTM

机译:使用Deep LSTM提高Punjabi语音识别系统的长语言依赖性的准确性

获取原文
获取原文并翻译 | 示例
       

摘要

Long short term memory (LSTM) is a powerful model in building of an ASR system whereas standard recurrent networks are generally inefficient to obtain better performance. Although these issues are addressed in LSTM neural network architecture but their performance get degraded on long contextual information. Recent experiments show that LSTM and their improved approaches like Deep LSTM requires a lot of tuning in training and experiences. In this paper Deep LSTM models are built on long contextual sentences by selecting optimal value of batch size, layer, and activation functions. It also indulge comparative study of train and test perplexity through computation of word error rate. Furthermore, we use hybrid discriminative approaches with different variants of iterations which shows significant improvement with Deep LSTM networks. Experiments are mainly perform on single sentences or one to two concatenated sentences. Deep LSTM achieves performance improvement of 3-4% over conventional Language Models (LMs) and modelling classifier approaches with acceptable word error rate on top of state-of-the-art Punjabi speech recognition system.
机译:长期短期内存(LSTM)是在ASR系统构建的强大模型,而标准的经常性网络通常效率低下以获得更好的性能。虽然这些问题是在LSTM神经网络架构中解决的,但它们的性能在长的上下文信息上变得劣化。最近的实验表明,LSTM及其改进的方法,如深入LSTM,需要大量调整培训和经验。在本文中,深入的LSTM模型是通过选择批量大小,层和激活功能的最佳值来构建长的上下文句子。它还通过计算单词误差率来沉迷于火车和测试困惑的比较研究。此外,我们使用具有不同迭代变体的混合鉴别方法,其具有深入的LSTM网络显着改善。实验主要是单句或一个到两个连接句子。深层LSTM在传统的语言模型(LMS)上实现了3-4%的性能提高,并在最先进的旁遮普语音识别系统顶部具有可接受的单词错误率的分类器方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号