首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >EIGENTONGUE FEATURE EXTRACTION FOR AN ULTRASOUND-BASED SILENT SPEECH INTERFACE
【24h】

EIGENTONGUE FEATURE EXTRACTION FOR AN ULTRASOUND-BASED SILENT SPEECH INTERFACE

机译:基于超声的无声语音接口的Eigentongue特征提取

获取原文
获取外文期刊封面目录资料

摘要

The article compares two approaches to the description of ultrasound vocal tract images for application in a "silent speech interface," one based on tongue contour modeling, and a second, global coding approach in which images are projected onto a feature space of Eigentongues. A curvature-based lip profile feature extraction method is also presented. Extracted visual features are input to a neural network which learns the relation between the vocal tract configuration and line spectrum frequencies (LSF) contained in a one-hour speech corpus. An examination of the quality of LSF's derived from the two approaches demonstrates that the eigentongues approach has a more efficient implementation and provides superior results based on a normalized mean squared error criterion.
机译:该物品将超声声道图像的描述进行比较,以便在基于舌片轮廓建模的“静音语音接口”中应用两个方法,以及将图像投射到尖端的特征空间中的第二种全局编码方法。 还提出了一种基于曲率的唇形轮廓特征提取方法。 提取的可视特征被输入到神经网络,该神经网络学习在一小时语音语料库中包含的声乐道配置和线频谱频率(LSF)之间的关系。 对来自两种方法的LSF质量的审查表明,Eigentongs方法具有更有效的实施,并基于归一化平均平方误差标准提供优越的结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号