首页> 外文会议>Asia-Pacific Signal and Information Processing Association Annual Summit and Conference >Multimodal Speech Driven Facial Shape Animation Using Deep Neural Networks
【24h】

Multimodal Speech Driven Facial Shape Animation Using Deep Neural Networks

机译:使用深度神经网络的多峰语音驱动面部形状动画

获取原文

摘要

In this paper we present a deep learning multimodal approach for speech driven generation of face animations. Training a speaker independent model, capable of generating different emotions of the speaker, is crucial for realistic animations. Unlike the previous approaches which either use acoustic features or phoneme label features to estimate the facial movements, we utilize both modalities to generate natural looking speaker independent lip animations synchronized with affective speech. A phoneme-based model qualifies generation of speaker independent animation, whereas an acoustic feature-based model enables capturing affective variation during the animation generation. We show that our multimodal approach not only performs significantly better on affective data, but improves performance over neutral data as well. We evaluate the proposed multimodal speech-driven animation model using two large scale datasets, GRID and SAVEE, by reporting the mean squared error (MSE) over various network structures.
机译:在本文中,我们为脸部动画的语音驱动产生了深度学习多模式方法。培训一个能够产生扬声器的不同情绪的扬声器独立模型对现实动画至关重要。与使用声学特征或音素标签的方法来估计面部运动的方法不同,我们使用两种方式生成与情感语音同步的自然观看的扬声器独立的唇动画。基于音素的模型符合扬声器独立动画的生成,而基于声学特征的模型可以在动画生成期间捕获有影响变化。我们表明,我们的多模式方法不仅对情感数据进行了明显更好的表现,而且还提高了中性数据的性能。我们通过在各种网络结构上报告平均平方误差(MSE)来评估使用两个大型数据集,网格和Savee来评估所提出的多模式语音驱动的动画模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号