首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >DYNAMIC FRAME SKIPPING FOR FAST SPEECH RECOGNITION IN RECURRENT NEURAL NETWORK BASED ACOUSTIC MODELS
【24h】

DYNAMIC FRAME SKIPPING FOR FAST SPEECH RECOGNITION IN RECURRENT NEURAL NETWORK BASED ACOUSTIC MODELS

机译:基于复发性神经网络声学模型的快速语音识别动态帧

获取原文

摘要

A recurrent neural network is a powerful tool for modeling sequential data such as text and speech. While recurrent neural networks have achieved record-breaking results in speech recognition, one remaining challenge is their slow processing speed. The main cause comes from the nature of recurrent neural networks that read only one frame at each time step. Therefore, reducing the number of reads is an effective approach to reducing processing time. In this paper, we propose a novel recurrent neural network architecture called Skip-RNN, which dynamically skips speech frames that are less important. The Skip-RNN consists of an acoustic model network and skip-policy network that are jointly trained to classify speech frames and determine how many frames to skip. We evaluate our proposed approach on the Wall Street Journal corpus and show that it can accelerate acoustic model computation by up to 2.4 times without any noticeable degradation in transcription accuracy.
机译:经常性神经网络是一种强大的工具,用于建模顺序数据,如文本和语音。虽然经常性的神经网络已经实现了语音识别的录制效果,但一个剩下的挑战是他们的慢速处理速度。主要原因来自于经常性神经网络的性质,每次只读一个帧。因此,减少读数的数量是减少处理时间的有效方法。在本文中,我们提出了一种名为Skip-RNN的新型复发性神经网络架构,其动态地跳过了不太重要的语音帧。 Skip-RNN由声学模型网络和跳过策略网络组成,这些网络是共同培训的,以分类语音帧并确定跳过多少帧。我们在华尔街日志语料库中评估了我们提出的方法,并表明它可以将声学模型计算加速到2.4倍,而没有任何明显的转录精度下降。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号