【24h】

Real-time speech-driven lip synchronization

机译:实时语音驱动的嘴唇同步

获取原文

摘要

Speech-driven lip synchronization, an important part of facial animation, is to animate a face model to render lip movements that are synchronized with the acoustic speech signal. It has many applications in human-computer interaction. In this paper, we present a framework that systematically addresses multimodal database collection and processing and real-time speech-driven lip synchronization using collaborative filtering which is a data-driven approach used by many online retailers to recommend products. Mel-frequency cepstral coefficients (MFCCs) with their delta and acceleration coefficients and Facial Animation Parameters (FAPs) supported by MPEG-4 for the visual representation of speech are utilized as acoustic features and animation parameters respectively. The proposed system is speaker independent and real-time capable. The subjective experiments show that the proposed approach generates a natural facial animation.
机译:语音驱动的嘴唇同步是面部动画的重要部分,它是对面部模型进行动画处理,以渲染与声学语音信号同步的嘴唇运动。它在人机交互中有许多应用。在本文中,我们提出了一个框架,该框架使用协作过滤系统地解决了多模式数据库的收集和处理以及实时语音驱动的口形同步,这是许多在线零售商推荐产品的数据驱动方法。带有频率增量和加速度系数的梅尔频率倒谱系数(MFCC)和MPEG-4支持的用于语音视觉表示的面部动画参数(FAP)被分别用作声学特征和动画参数。所提出的系统是独立于说话者的并且具有实时能力。主观实验表明,所提出的方法产生了自然的面部动画。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号