首页> 外文会议>Annual conference of the International Speech Communication Association >Pipelined Back-Propagation for Context-Dependent Deep Neural Networks
【24h】

Pipelined Back-Propagation for Context-Dependent Deep Neural Networks

机译:管制依赖性深神经网络的流水线反向传播

获取原文

摘要

The Context-Dependent Deep-Neural-Network HMM, or CD DNN-HMM, is a recently proposed acoustic-modeling tech nique for HMM-based speech recognition that can greatly out perform conventional Gaussian-mixture based HMMs. For ex ample, a CD-DNN-HMM trained on the 2000h Fisher corpus achieves 14.4% word error rate on the Hub5'00-FSH speaker independent phone-call transcription task, compared to 19.6% obtained by a state-of-the-art, conventional discriminatively trained GMM-based HMM. That CD-DNN-HMM, however, took 59 days to train on a modern GPGPU-the immense computational cost of the mini batch based back-propagation (BP) training is a major road block. Unlike the familiar Baum-Welch training for conven tional HMMs, BP cannot be efficiently parallelized across data. In this paper we show that the pipelined approximation to BP, which parallelizes computation with respect to layers, is an efficient way of utilizing multiple GPGPU cards in a single server. Using 2 and 4 GPGPUs, we achieve a 1.9 and 3.3 times end-to-end speed-up, at parallelization efficiency of 0.95 and 0.82, respectively, at no loss of recognition accuracy.
机译:依赖上下文的深神经网络HMM,或CD DNN-HMM是最近提出的基于赫姆的语音识别的声学建模技术Nique,其可以大大地执行基于传统的高斯混合的HMM。对于EX充足,在2000h Fisher语料库上培训的CD-DNN-HMM在HUB5'00-FSH扬声器独立的电话呼叫转录任务上实现了14.4%的字错误率,而19.6%通过─艺术,常规的基于GMM的MMM训练训练。然而,CD-DNN-HMM培训了现代GPGPU的59天 - 迷你批次的背部传播(BP)培训的巨大计算成本是一个主要的道路块。与常规HMMS的熟悉的BAUM-Welch训练不同,BP不能跨数据有效地平行化。在本文中,我们示出了向BP的流水线近似,该BP是相对于层的计算,是利用单个服务器中多个GPGPU卡的有效方式。使用2和4 GPGPU,我们分别在0.95和0.82的并行化效率下实现1.9和3.3倍的端到端加速,无识别精度。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号