首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >Focus on the Present: A Regularization Method for the ASR Source-Target Attention Layer
【24h】

Focus on the Present: A Regularization Method for the ASR Source-Target Attention Layer

机译:专注于现在:ASR源极关注层的正则化方法

获取原文

摘要

This paper introduces a novel method to diagnose the source-target attention in state-of-the-art end-to-end speech recognition models with joint connectionist temporal classification (CTC) and attention training. Our method is based on the fact that both, CTC and source-target attention, are acting on the same encoder representations. To understand the functionality of the attention, CTC is applied to compute the token posteriors given the attention outputs. We found that the source-target attention heads are able to predict several tokens ahead of the current one. Inspired by the observation, a new regularization method is proposed which leverages CTC to make source-target attention more focused on the frames corresponding to the output token being predicted by the decoder. Experiments reveal stable improvements up to 7% and 13% relatively with the proposed regularization on TED-LIUM 2 and Librispeech.
机译:本文介绍了一种新的方法,可以在最先进的端到端语音识别模型中诊断源极关注,具有关节连接主人时间分类(CTC)和注意力训练。 我们的方法基于以下事实,即CTC和源极关注,采用相同的编码器表示。 要了解注意力,CTC应用于指定注意输出来计算令牌后辅导。 我们发现,源极点关注头能够在当前的位置预测几个令牌。 通过观察的启发,提出了一种新的正则化方法,其利用CTC使源极关注更专注于对应于解码器预测的输出令牌的帧。 实验显示稳定的改善高达7%和13%,在TED-lium 2和Libriispeech上的拟议正则化。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号