首页> 外文会议>2018 IEEE Spoken Language Technology Workshop >Improving Attention-Based End-to-End ASR Systems with Sequence-Based Loss Functions
【24h】

Improving Attention-Based End-to-End ASR Systems with Sequence-Based Loss Functions

机译:使用基于序列的损失函数改善基于注意力的端到端ASR系统

获取原文
获取原文并翻译 | 示例

摘要

Acoustic model and language model (LM) have been two major components in conventional speech recognition systems. They are normally trained independently, but recently there has been a trend to optimize both components simultaneously in a unified end-to-end (E2E) framework. However, the performance gap between the E2E systems and the traditional hybrid systems suggests that some knowledge has not yet been fully utilized in the new framework. An observation is that the current attention-based E2E systems could produce better recognition results when decoded with LMs which are independently trained with the same resource. In this paper, we focus on how to improve attention-based E2E systems without increasing model complexity or resorting to extra data. A novel training strategy is proposed for multi-task training with the connectionist temporal classification (CTC) loss. The sequence-based minimum Bayes risk (MBR) loss is also investigated. Our experiments on SWB 300hrs showed that both loss functions could significantly improve the baseline model performance. The additional gain from joint-LM decoding remains the same for CTC trained model but is only marginal for MBR trained model. This implies that while CTC loss function is able to capture more acoustic knowledge, MBR loss function exploits more word/character dependency.
机译:声学模型和语言模型(LM)已成为常规语音识别系统中的两个主要组件。它们通常是独立培训的,但是最近出现了在统一的端到端(E2E)框架中同时优化两个组件的趋势。但是,端到端系统与传统混合系统之间的性能差距表明,一些知识尚未在新框架中得到充分利用。可以观察到,当前的基于注意力的端到端系统在使用由相同资源独立训练的LM进行解码时,可以产生更好的识别结果。在本文中,我们专注于如何在不增加模型复杂性或不诉诸额外数据的情况下改善基于注意力的端到端系统。提出了一种新颖的训练策略,该方法用于在连接者时间分类(CTC)丢失的情况下进行多任务训练。还研究了基于序列的最小贝叶斯风险(MBR)损失。我们在SWB 300hrs上进行的实验表明,两种损失函数都可以显着改善基线模型的性能。对于CTC训练的模型,来自联合LM解码的额外增益保持不变,但对于MBR训练的模型而言,仅是很小的。这意味着,尽管CTC损失函数能够捕获更多的声学知识,但MBR损失函数却利用了更多的单词/字符依赖性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号