首页> 外文会议>Chinese Conference on biometric recognition >Prioritized Grid Highway Long Short-Term Memory-Based Universal Background Model for Speaker Verification
【24h】

Prioritized Grid Highway Long Short-Term Memory-Based Universal Background Model for Speaker Verification

机译:优先基于网格公路长期短期记忆的说话人验证通用背景模型

获取原文

摘要

Prioritized grid long short-term memory (pGLSTM) has been shown to improve automatic speech recognition efficiently. In this paper, we implement this state-of-the-art model of ASR tasks for text-independent Chinese language speaker verification tasks in which DNN/i-Vector (DNN-based i-Vector) framework is adopted along with PLDA backend. To fully explore the performance, we compared the presented pGLSTM based UBM to GMM-UBM and HLSTM-UBM. Due to constraint of the amount of Chinese transcribed corpus for ASR training, we also explore an adaptation method by firstly training the pGLSTM-UBM on English language with large amount of corpus and use a PLDA adaptation backend to fit into Chinese language before the final speaker verification scoring. Experiments show that both pGLSTM-UBM model with corresponding PLDA backend and pGLSTM-UBM with adapted PLDA backend achieve better performance than the traditional GMM-UBM model. Additionally the pGLSTM-UBM with PLDA backend achieves performance of 4.94% EER in 5 s short utterance and 1.97% EER in 10 s short utterance, achieving 47% and 51% drop comparing to that of GMM. Experiment results imply that DNN from ASR tasks can expand the advantage of UBM model especially in short utterance and that better DNN model for ASR tasks could achieve extra gain in speaker verification tasks.
机译:优先的网格长短期记忆(pGLSTM)已被证明可以有效地改善自动语音识别。在本文中,我们将DSR / i-Vector(基于DNN的i-Vector)框架与PLDA后端一起采用,从而针对与文本无关的中文说话者验证任务实现了这种ASR任务的最新模型。为了全面研究性能,我们将提出的基于pGLSTM的UBM与GMM-UBM和HLSTM-UBM进行了比较。由于用于ASR训练的中文转录语料库数量的限制,我们还探索了一种适应方法,首先在具有大量语料库的英语上训练pGLSTM-UBM,然后在最终讲者面前使用PLDA适应后端来适应中文验证评分。实验表明,带有相应PLDA后端的pGLSTM-UBM模型和带有适配PLDA后端的pGLSTM-UBM均比传统的GMM-UBM模型具有更好的性能。此外,带有PLDA后端的pGLSTM-UBM在5秒钟的短时间内可实现4.94%的EER,在10秒钟的短时间内可实现1.97%的EER,与GMM相比,可实现47%和51%的下降。实验结果表明,来自ASR任务的DNN可以扩展UBM模型的优势,特别是在短话语中,而针对ASR任务的更好的DNN模型可以在说话者验证任务中获得额外的收益。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号