首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >Using contextual information in joint factor eigenspace MLLR for speech recognition in diverse scenarios
【24h】

Using contextual information in joint factor eigenspace MLLR for speech recognition in diverse scenarios

机译:在不同场景下使用联合因子本征空间MLLR中的上下文信息进行语音识别

获取原文

摘要

This paper presents a new approach for rapid adaptation in the presence of highly diverse scenarios that takes advantage of information describing the input signals. We introduce a new method for joint factorisation of the background and the speaker in an eigenspace MLLR framework: Joint Factor Eigenspace MLLR (JFEMLLR). We further propose to use contextual information describing the speaker and background, such as tags or more complex metadata, to provide an immediate estimation of the best MLLR transformation for the utterance. This provides instant adaptation, since it does not require any transcription from a previous decoding stage. Evaluation in a highly diverse Automatic Speech Recognition (ASR) task, a modified version of WSJCAM0, yields an improvement of 26.9% over the baseline, which is an extra 1.2% reduction over two-pass MLLR adaptation.
机译:本文提出了一种在高度多样化的情况下快速适应的新方法,该方法利用了描述输入信号的信息。我们介绍了一种在本征空间MLLR框架中对背景和说话人进行联合分解的新方法:联合因子本征空间MLLR(JFEMLLR)。我们进一步建议使用描述说话者和背景的上下文信息(例如标签或更复杂的元数据)来为言语提供最佳MLLR转换的立即估计。这提供了即时适应性,因为它不需要来自先前解码阶段的任何转录。在高度多样化的自动语音识别(ASR)任务(WSJCAM0的修改版)中进行评估,比基线提高了26.9%,比两次通过MLLR自适应降低了1.2%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号