首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >Teach an All-rounder with Experts in Different Domains
【24h】

Teach an All-rounder with Experts in Different Domains

机译:在不同领域的专家教授一个全方位的人

获取原文

摘要

In many automatic speech recognition (ASR) tasks, an ideal model has to be applicable over multiple domains. In this paper, we propose to teach an all-rounder with experts in different domains. Concretely, we build a multi-domain acoustic model by applying the teacher-student training framework. First, for each domain, a teacher model (domain-dependent model) is trained by fine-tuning a multi-condition model with domain-specific subset. Then all these teacher models are used to teach one single student model simultaneously. We perform experiments on two predefined domain setups. One is domains with different speaking styles, the other is near-field, far-field and far-field with noise. Moreover, two types of models are examined: deep feedforward sequential memory network (DFSMN) and long short term memory (LSTM). Experimental results show that the model trained with this framework outperforms not only multi-condition model but also domain-dependent model. Specially, our training method provides up to 10.4% relative character error rate improvement over baseline model (multi-condition model).
机译:在许多自动语音识别(ASR)任务中,一个理想的模型必须适用于多个域。在本文中,我们建议在不同领域的专家教授一位全方位的专家。具体地,我们通过应用师生培训框架来构建多域声学模型。首先,对于每个域,通过微调特定于域子集的多条件模型来训练教师模型(域依赖模型)。然后所有这些教师模型都用于同时教授一个单一学生模型。我们在两个预定义的域设置上执行实验。一个是具有不同讲话方式的域,另一个是近场,远场和远场,噪音。此外,检查了两种类型的模型:深馈通顺序存储器网络(DFSMN)和长短期内存(LSTM)。实验结果表明,该框架训练的模型不仅占据了多条件模型,还涉及域依赖模型。特别是,我们的训练方法提供了基线模型(多条件模型)上的高达10.4%的相对字符错误率改进。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号