首页> 外文期刊>Audio, Speech, and Language Processing, IEEE Transactions on >Maximum Entropy-Based Reinforcement Learning Using a Confidence Measure in Speech Recognition for Telephone Speech
【24h】

Maximum Entropy-Based Reinforcement Learning Using a Confidence Measure in Speech Recognition for Telephone Speech

机译:电话语音识别中基于置信度的最大熵增强学习

获取原文
获取原文并翻译 | 示例
           

摘要

In this paper, a novel confidence-based reinforcement learning (RL) scheme to correct observation log-likelihoods and to address the problem of unsupervised compensation with limited estimation data is proposed. A two-step Viterbi decoding is presented which estimates a correction factor for the observation log-likelihoods that makes the recognized and neighboring HMMs more or less likely by using a confidence score. If regions in the output delivered by the recognizer exhibit low confidence scores, the second Viterbi decoding will tend to focus the search on neighboring models. In contrast, if recognized regions exhibit high confidence scores, the second Viterbi decoding will tend to retain the recognition output obtained at the first step. The proposed RL mechanism is modeled as the linear combination of two metrics or information sources: the acoustic model log-likelihood and the logarithm of a confidence metric. A criterion based on incremental conditional entropy maximization to optimize a linear combination of metrics or information sources online is also presented. The method requires only one utterance, as short as 0.7 s, and can lead to significant reductions in word error rate (WER) between 3% and 18%, depending on the task, training-testing conditions, and method used to optimize the proposed RL scheme. In contrast to ordinary feature compensation and model parameter adaptation methods, the confidence-based RL method takes place in the frame log-likelihood domain. Consequently, as shown in the results presented here, it is complementary to feature compensation and to model adaptation techniques.
机译:本文提出了一种新的基于置信度的强化学习(RL)方案,用于纠正观测对数似然率并解决估计数据有限的无监督补偿问题。提出了两步维特比解码,该维特比解码通过使用置信度分数来估计观察对数似然的校正因子,该校正因子使识别出的HMM和相邻HMM或多或少地具有可能性。如果识别器传递的输出中的区域表现出较低的置信度,则第二维特比解码将趋向于将搜索集中在相邻模型上。相反,如果识别的区域表现出高置信度分数,则第二维特比解码将倾向于保留在第一步获得的识别输出。所提出的RL机制被建模为两个度量或信息源的线性组合:声学模型的对数似然性和置信度度量的对数。还提出了一种基于增量条件熵最大化的指标,用于在线优化指标或信息源的线性组合。该方法仅需一次话语,短至0.7 s,并且可以根据任务,训练测试条件以及用于优化建议方案的方法,将字错误率(WER)显着降低3%至18%。 RL方案。与普通特征补偿和模型参数自适应方法相反,基于置信度的RL方法在帧对数似然域中进行。因此,如此处呈现的结果所示,它是特征补偿和模型自适应技术的补充。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号