首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >Extended low-rank plus diagonal adaptation for deep and recurrent neural networks
【24h】

Extended low-rank plus diagonal adaptation for deep and recurrent neural networks

机译:用于深度和递归神经网络的扩展低秩加对角线适应

获取原文

摘要

Recently, the low-rank plus diagonal (LRPD) adaptation was proposed for speaker adaptation of deep neural network (DNN) models. The LRPD restructures the adaptation matrix as a superposition of a diagonal matrix and a product of two low-rank matrices. In this paper, we extend the LRPD adaptation into the subspace-based approach to further reduce the speaker-dependent (SD) footprint. We apply the extended LRPD (eLRPD) adaptation for the DNN and LSTM models with emphasis placed on the applicability of the adaptation to large-scale speech recognition systems. To speed up the adaptation in test time, we propose the bottleneck (BN) caching approach to eliminate the redundant computations during multiple sweeps of development data. Experimental results on the short message dictation (SMD) task show that the eLRPD adaptation can reduce the SD footprints by 82% for the SVD DNN and 96% for the LSTM-RNN over the linear adaptation, while maintaining the comparable accuracy. The BN caching achieves up to 3.5 times speedup in adaptation at no loss of recognition accuracy.
机译:最近,提出了低秩加对角线(LRPD)自适应用于深度神经网络(DNN)模型的说话者自适应。 LRPD将适应矩阵重构为对角矩阵和两个低秩矩阵的乘积的叠加。在本文中,我们将LRPD适应性扩展到基于子空间的方法中,以进一步减少说话者相关(SD)的占用空间。我们将扩展的LRPD(eLRPD)适应性应用于DNN和LSTM模型,重点放在适应性在大型语音识别系统上的适用性。为了加快测试时间的适应性,我们提出了瓶颈(BN)缓存方法,以消除多次扫描开发数据期间的冗余计算。短消息听写(SMD)任务的实验结果表明,与线性自适应相比,eLRPD自适应可以使SVD DNN和LSTM-RNN的SD占用空间减少82 \%,而LSTM-RNN可以减少96%。 BN缓存的自适应速度提高了3.5倍,而不会损失识别精度。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号