首页> 外文会议> >Articulatory Feature-Based Methods for Acoustic and Audio-Visual Speech Recognition: Summary from the 2006 JHU Summer workshop
【24h】

Articulatory Feature-Based Methods for Acoustic and Audio-Visual Speech Recognition: Summary from the 2006 JHU Summer workshop

机译:基于发音特征的声音和视听语音识别方法:2006年JHU夏季研讨会的总结

获取原文
获取外文期刊封面目录资料

摘要

We report on investigations, conducted at the 2006 Johns Hopkins Workshop, into the use of articulatory features (AFs) for observation and pronunciation models in speech recognition. In the area of observation modeling, we use the outputs of AF classifiers both directly, in an extension of hybrid HMMeural network models, and as part of the observation vector, an extension of the "tandem" approach. In the area of pronunciation modeling, we investigate a model having multiple streams of AF states with soft synchrony constraints, for both audio-only and audio-visual recognition. The models are implemented as dynamic Bayesian networks, and tested on tasks from the small-vocabulary switchboard (SVitchboard) corpus and the CUAVE audio-visual digits corpus. Finally, we analyze AF classification and forced alignment using a newly collected set of feature-level manual transcriptions
机译:我们报告了在2006年约翰·霍普金斯讲习班上进行的有关在语音识别中将发音特征(AF)用于观察和发音模型的调查。在观察建模领域,我们在混合HMM /神经网络模型的扩展中直接使用了AF分类器的输出,并且在“观察”向量的一部分中,直接使用了“串联”方法的扩展。在语音建模领域,我们研究了具有多个具有软同步约束的AF状态流的模型,用于纯音频和视听识别。这些模型被实现为动态贝叶斯网络,并在小型词汇总机(SVitchboard)语料库和CUAVE视听数字语料库的任务上进行了测试。最后,我们使用一组新收集的功能级别手动转录来分析AF分类和强制对齐

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号