首页> 外文会议>International Conference on Statistical Language and Speech Processing >Merging of Native and Non-native Speech for Low-resource Accented ASR
【24h】

Merging of Native and Non-native Speech for Low-resource Accented ASR

机译:融合低资源突出的ASR的本机和非原生语音

获取原文

摘要

This paper presents our recent study on low-resource automatic speech recognition (ASR) system with accented speech. We propose multi-accent Subspace Gaussian Mixture Models (SGMM) and accent-specific Deep Neural Networks (DNN) for improving non-native ASR performance. In the SGMM framework, we present an original language weighting strategy to merge the globally shared parameters of two models based on native and non-native speech respectively. In the DNN framework, a native deep neural net is fine-tuned to non-native speech. Over the non-native baseline, we achieved relative improvement of 15% for multi-accent SGMM and 34% for accent-specific DNN with speaker adaptation.
机译:本文介绍了我们最近关于具有重音语音的低资源自动语音识别(ASR)系统的研究。我们提出了多口语子空间高斯混合模型(SGMM)和强调特定的深神经网络(DNN),用于改善非本机ASR性能。在SGMM框架中,我们提出了一种原始语言加权策略,以分别基于本机和非原生言论合并两个模型的全局共享参数。在DNN框架中,原生深神经网络被微调到非原生语音。在非本地基线上,我们实现了多口径SGMM的相对提高15%,对于具有扬声器适应的口音DNN,34%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号