首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >Exploring A Zero-Order Direct Hmm Based on Latent Attention for Automatic Speech Recognition
【24h】

Exploring A Zero-Order Direct Hmm Based on Latent Attention for Automatic Speech Recognition

机译:基于自动语音识别的潜在关注探索零级直接嗯

获取原文
获取外文期刊封面目录资料

摘要

In this paper, we study a simple yet elegant latent variable attention model for automatic speech recognition (ASR) which enables an integration of attention sequence modeling into the direct hidden Markov model (HMM) concept. We use a sequence of hidden variables that establishes a mapping from output labels to input frames. Inspired by the direct HMM model, we assume a decomposition of the label sequence posterior into emission and transition probabilities using zero-order assumption and incorporate both Transformer and LSTM attention models into it. The method keeps the explicit alignment as part of the stochastic model and combines the ease of the end-to-end training of the attention model as well as an efficient and simple beam search. To study the effect of the latent model, we qualitatively analyze the alignment behavior of the different approaches. Our experiments on three ASR tasks show promising results in WER with more focused alignments in comparison to the attention models.
机译:在本文中,我们研究了一个简单而优雅的潜在可变注意模型,用于自动语音识别(ASR),它能够将注意序列建模集成到直接隐藏的马尔可夫模型(HMM)概念中。我们使用一系列隐藏变量,该序列将从输出标签建立映射到输入帧。灵感来自直接肝脏模型,我们假设使用零阶假设的发射和过渡概率后面的标签序列分解,并将变压器和LSTM注意模型结合到其中。该方法将显式对齐作为随机模型的一部分,并结合了注意力模型的易于结束训练以及高效和简单的光束搜索。为研究潜伏模型的影响,我们定性地分析了不同方法的对准行为。我们对三个ASR任务的实验表明,与注意模型相比,WER的有希望的结果具有更大的对齐。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号