首页> 外文期刊>Journal of VLSI signal processing systems for signal, image, and video technology >Improving the Decoding Efficiency of Deep Neural Network Acoustic Models by Cluster-Based Senone Selection
【24h】

Improving the Decoding Efficiency of Deep Neural Network Acoustic Models by Cluster-Based Senone Selection

机译:通过基于簇的Senone选择提高深层神经网络声学模型的解码效率

获取原文
获取原文并翻译 | 示例
获取外文期刊封面目录资料

摘要

In this paper, we propose a cluster-based senone selection method to speed up the computation of deep neural networks (DNN) at the decoding time of automatic speech recognition (ASR) systems. In DNN-based acoustic models, the large number of senones at the output layer is one of the main causes that lead to the high computation complexity of DNNs. Inspired by the mixture selection method designed for the Gaussian mixture model (GMM)-based acoustic models, only a subset of the senones at the output layer of DNNs are selected to calculate the posterior probabilities in our proposed method. The senone selection strategy is derived by clustering acoustic features according to their transformed representations at the top hidden layer of the DNN acoustic model. Experimental results on Mandarin speech recognition tasks show that the average number of DNN parameters used for computation can be reduced by 22% and the overall speed of the recognition process can be accelerated by 13% without significant performance degradation after using our proposed method. Experimental results on the Switchboard task demonstrate that our proposed method can reduce the average number of DNN parameters used for computation by 38.8% for conventional DNN modeling and 22.7% for low-rank DNN modeling respectively with negligible performance loss.
机译:在本文中,我们提出了一种基于簇的senone选择方法,以在自动语音识别(ASR)系统的解码时加速深度神经网络(DNN)的计算。在基于DNN的声学模型中,输出层中大量的senone是导致DNN计算复杂性高的主要原因之一。受针对基于高斯混合模型(GMM)的声学模型设计的混合选择方法的启发,在我们提出的方法中,仅选择了DNN输出层中的senone的子集来计算后验概率。 senone选择策略是通过根据DNN声学模型的顶部隐藏层上的变换特征对声学特征进行聚类而得出的。对普通话语音识别任务的实验结果表明,使用我们提出的方法后,用于计算的DNN参数的平均数量可以减少22%,识别过程的整体速度可以提高13%,而不会显着降低性能。在Switchboard任务上的实验结果表明,我们提出的方法可以将用于计算的DNN参数的平均数量分别减少常规DNN建模的38.8%和低秩DNN建模的22.7%,而性能损失可忽略不计。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号