首页> 外文会议> >Speaker-independent speech-driven facial animation using a hierarchical model
【24h】

Speaker-independent speech-driven facial animation using a hierarchical model

机译:使用分层模型的与说话者无关的语音驱动面部动画

获取原文

摘要

We present a system capable of producing video-realistic videos of a speaker given audio only. The audio input signal requires no phonetic labelling and is speaker independent. The system requires only a small training set of video to achieve convincing realistic facial synthesis. The system learns the natural mouth and face dynamics of a speaker to allow new facial poses, unseen in the training video, to be synthesised. To achieve this we have developed a novel approach which utilises a hierarchical and nonlinear PCA model which couples speech and appearance. We show that the model is capable of synthesising videos of a speaker using new audio segments from both previously heard and unheard speakers. The model is highly compact making it suitable for a wide range of real-time applications in multimedia and telecommunications using standard hardware.
机译:我们提出了一个能够产生给定音频的扬声器的视频现实视频的系统。音频输入信号不需要语音标记,并且是扬声器独立的。该系统只需要一组小型训练视频来实现令人信服的现实面部合成。该系统学习扬声器的自然口和面部动态,以允许新的面部姿势,在训练视频中进行看不见。为实现这一点,我们开发了一种新颖的方法,它利用耦合语音和外观的分层和非线性PCA模型。我们表明该模型能够使用先前听说和闻名扬声器的新音频段来合成扬声器的视频。该模型非常紧凑,适用于使用标准硬件的多媒体和电信中的各种实时应用。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号