首页> 外文OA文献 >Robust speech recognition based on the second-order difference cochlear model
【2h】

Robust speech recognition based on the second-order difference cochlear model

机译:基于二阶差分耳蜗模型的鲁棒语音识别

摘要

MFCC (Mel-Frequency Cepstral Coefficients) is a kind of traditional speech feature widely used in speech recognition. The error rate of the speech recognition algorithm using MFCC and CDHMM is known to be very low in a clean speech environment, but it increases greatly in a noisy environment, especially in the white noisy environment. We propose a new kind of speech feature called the auditory spectrum based feature (ASBF) that is based on the second-order difference cochlear model of the human auditory system. This new speech feature can track the speech formants and the selection scheme of this feature is based on both the second-order difference cochlear model and primary auditory nerve processing model of the human auditory system. In our experiment, the performance of MFCC and ASBF are compared in both clean and noisy environments when left-to-right CDHMM with 6 states and 5 Gaussian mixtures is used. The experimental result shows that the ASBF is much more robust to noise than MFCC. When only 5 frequency components are used in ASBF, the error rate is approximately 38% lower than the traditional MFCC with 39 parameters in the condition of S/N=10 dB with white noise
机译:MFCC(Mel-频率倒谱系数)是一种广泛用于语音识别的传统语音功能。众所周知,在干净的语音环境中,使用MFCC和CDHMM的语音识别算法的错误率非常低,但是在嘈杂的环境中,尤其是在白色嘈杂的环境中,其错误率会大大增加。我们提出了一种新的语音特征,称为基于听觉频谱的特征(ASBF),它是基于人类听觉系统的二阶差异耳蜗模型。这个新的语音特征可以跟踪语音共振峰,并且该特征的选择方案基于人类听觉系统的二阶差分耳蜗模型和主听神经处理模型。在我们的实验中,当使用具有6种状态和5种高斯混合的左右CDHMM时,在干净和嘈杂的环境中比较了MFCC和ASBF的性能。实验结果表明,ASBF的抗噪声能力比MFCC强得多。如果在S / N = 10 dB和白噪声的情况下,在ASBF中仅使用5个频率分量,则误码率将比具有39个参数的传统MFCC降低38%。

著录项

  • 作者

    Wan WG; Au OC;

  • 作者单位
  • 年度 2001
  • 总页数
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类

相似文献

  • 外文文献
  • 中文文献
  • 专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号