首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >Convolutional, Long Short-Term Memory, fully connected Deep Neural Networks
【24h】

Convolutional, Long Short-Term Memory, fully connected Deep Neural Networks

机译:卷积,长短期记忆,完全连接的深度神经网络

获取原文

摘要

Both Convolutional Neural Networks (CNNs) and Long Short-Term Memory (LSTM) have shown improvements over Deep Neural Networks (DNNs) across a wide variety of speech recognition tasks. CNNs, LSTMs and DNNs are complementary in their modeling capabilities, as CNNs are good at reducing frequency variations, LSTMs are good at temporal modeling, and DNNs are appropriate for mapping features to a more separable space. In this paper, we take advantage of the complementarity of CNNs, LSTMs and DNNs by combining them into one unified architecture. We explore the proposed architecture, which we call CLDNN, on a variety of large vocabulary tasks, varying from 200 to 2,000 hours. We find that the CLDNN provides a 4–6% relative improvement in WER over an LSTM, the strongest of the three individual models.
机译:卷积神经网络(CNN)和长短期记忆(LSTM)都已在各种语音识别任务中显示出对深度神经网络(DNN)的改进。 CNN,LSTM和DNN在建模能力上是互补的,因为CNN擅长减少频率变化,LSTM擅长时间建模,而DNN适合将要素映射到更可分离的空间。在本文中,我们通过将CNN,LSTM和DNN组合成一个统一的架构来利用它们的互补性。我们探索了建议的体系结构,我们将其称为CLDNN,用于各种大型词汇任务,时间从200到2,000小时不等。我们发现,相对于LSTM(三个模型中最强的模型),CLDNN的WER相对提高了4–6%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号