首页> 外文期刊>IEEE Transactions on Speech and Audio Proceeding >Maximum likelihood and minimum classification error factor analysis for automatic speech recognition
【24h】

Maximum likelihood and minimum classification error factor analysis for automatic speech recognition

机译:自动语音识别的最大似然和最小分类误差因子分析

获取原文
获取原文并翻译 | 示例
       

摘要

Hidden Markov models (HMMs) for automatic speech recognition rely on high dimensional feature vectors to summarize the short-time properties of speech. Correlations between features can arise when the speech signal is nonstationary or corrupted by noise. We investigate how to model these correlations using factor analysis, a statistical method for dimensionality reduction. Factor analysis uses a small number of parameters to model the covariance structure of high dimensional data. These parameters can be chosen in two ways: (1) to maximize the likelihood of observed speech signals, or (2) to minimize the number of classification errors. We derive an expectation-maximization (EM) algorithm for maximum likelihood estimation and a gradient descent algorithm for improved class discrimination. Speech recognizers are evaluated on two tasks, one small-sized vocabulary (connected alpha-digits) and one medium-sized vocabulary (New Jersey town names). We find that modeling feature correlations by factor analysis leads to significantly increased likelihoods and word accuracies. Moreover, the rate of improvement with model size often exceeds that observed in conventional HMM's.
机译:用于自动语音识别的隐马尔可夫模型(HMM)依赖于高维特征向量来总结语音的短时属性。当语音信号不稳定或被噪声破坏时,可能会出现功能之间的相关性。我们研究如何使用因子分析(降维的一种统计方法)对这些相关性进行建模。因子分析使用少量参数来建模高维数据的协方差结构。可以通过两种方式选择这些参数:(1)最大化观察到的语音信号的可能性,或(2)最小化分类错误的数量。我们推导了用于最大似然估计的期望最大化(EM)算法和用于改进类别识别的梯度下降算法。对语音识别器进行两项任务评估,一种是小型词汇(连接的字母数字),另一种是中型词汇(新泽西州的城镇名称)。我们发现,通过因素分析来建模特征相关性会导致可能性和单词准确性显着提高。此外,模型尺寸的改进速度通常超过传统HMM中观察到的速度。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号