首页> 外文会议>International Conference on intelligent science and big data engineering >Covariance Based Deep Feature for Text-Dependent Speaker Verification
【24h】

Covariance Based Deep Feature for Text-Dependent Speaker Verification

机译:基于协方差的深度特征,用于文本相关的说话人验证

获取原文
获取外文期刊封面目录资料

摘要

d-vector approach achieved impressive results in speaker verification. Representation is obtained at utterance level by calculating the mean of the frame level outputs of a hidden layer of the DNN. Although mean based speaker identity representation has achieved good performance, it ignores the variability of frames across the whole utterance, which consequently leads to information loss. This is particularly serious for text-dependent speaker verification, where within-utterance feature variability better reflects text variability than the mean. To address this issue, a new covariance based speaker representation is proposed in this paper. Here, covariance of the frame level outputs is calculated and incorporated into the speaker identity representation. The proposed approach is investigated within a joint multi-task learning framework for text-dependent speaker verification. Experiments on RSR2015 and RedDots showed that, covariance based deep feature can significantly improve the performance compared to the traditional mean based deep features.
机译:d矢量方法在说话人验证中取得了令人印象深刻的结果。通过计算DNN隐藏层的帧级别输出的平均值,以发声级别获得表示。尽管基于均值的说话人身份表示已经取得了良好的性能,但它忽略了整个发声中帧的可变性,从而导致信息丢失。这对于依赖文本的说话人验证尤为严重,其中话语内特征的变异性比均值更好地反映了文本变异性。为了解决这个问题,本文提出了一种新的基于协方差的说话人表示方法。在此,计算帧电平输出的协方差并将其并入说话者身份表示中。在联合多任务学习框架内对提出的方法进行了研究,以进行与文本相关的说话人验证。在RSR2015和RedDots上进行的实验表明,与传统的基于均值的深度特征相比,基于协方差的深度特征可以显着提高性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号