首页> 外文会议> >Robust face feature analysis for automatic speechreading and character animation
【24h】

Robust face feature analysis for automatic speechreading and character animation

机译:强大的人脸特征分析功能,可自动进行语音朗读和角色动画制作

获取原文

摘要

The robust acquisition of facial features needed for visual speech processing is fraught with difficulties which greatly increase the complexity of the machine vision system. This system must extract the inner lip contour from facial images with variations in pose, lighting, and facial hair. This paper describes a face feature acquisition system with robust performance in the presence of extreme lighting variations and moderate variations in pose. Furthermore, system performance is not degraded by facial hair or glasses. To find the position of a face reliably we search the whole image for facial features. These features are then combined and tests are applied, to determine whether any such combination actually belongs to a face. In order to find where the lips are, other features of the face, such as the eyes, must be located as well. Without this information it is difficult to reliably find the mouth in a complex image. Just the mouth by itself is easily missed or other elements in the image can be mistaken for a mouth. If camera position can be constrained to allow the nostrils to be viewed, then nostril tracking is used to both reduce computation and provide additional robustness. Once the nostrils are tracked from frame to frame using a tracking window the mouth area can be isolated and normalized for scale and rotation. A mouth detail analysis procedure is then used to estimate the inner lip contour and teeth and tongue regions. The inner lip contour and head movements are then mapped to synthetic face parameters to generate a graphical talking head synchronized with the original human voice. This information can also be used as the basis for visual speech features in an automatic speechreading system. Similar features were used in our previous automatic speechreading systems.
机译:视觉语音处理所需的面部特征的鲁棒性获取充满困难,这极大地增加了机器视觉系统的复杂性。该系统必须从面部图像中提取具有姿势,照明和面部毛发变化的内唇轮廓。本文描述了一种面部特征采集系统,该系统在极端光照变化和姿势适度变化的情况下具有强大的性能。此外,面部毛发或眼镜不会降低系统性能。为了可靠地找到面部的位置,我们在整个图像中搜索面部特征。然后将这些特征进行组合并进行测试,以确定任何此类组合是否实际上属于人脸。为了找到嘴唇的位置,还必须确定脸部的其他特征,例如眼睛。没有这些信息,就很难在复杂图像中可靠地找到嘴巴。仅嘴本身很容易被遗漏,或者图像中的其他元素可能被误认为嘴。如果可以限制摄像机位置以查看鼻孔,则可以使用鼻孔跟踪来减少计算量并提供额外的鲁棒性。一旦使用跟踪窗口逐帧跟踪鼻孔,就可以隔离口部区域并进行标准化以进行缩放和旋转。然后使用嘴部细节分析程序来估计内部嘴唇轮廓以及牙齿和舌头区域。然后将内部嘴唇轮廓和头部运动映射到合成人脸参数,以生成与原始人声同步的图形说话头。此信息也可以用作自动语音阅读系统中视觉语音功能的基础。我们以前的自动语音阅读系统使用了类似的功能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号