首页> 外文会议>IEEE Conference on Computer Vision and Pattern Recognition Workshops >Speech-Driven 3D Facial Animation with Implicit Emotional Awareness: A Deep Learning Approach
【24h】

Speech-Driven 3D Facial Animation with Implicit Emotional Awareness: A Deep Learning Approach

机译:具有内隐情感意识的语音驱动3D面部动画:一种深度学习方法

获取原文

摘要

We introduce a long short-term memory recurrent neural network (LSTM-RNN) approach for real-time facial animation, which automatically estimates head rotation and facial action unit activations of a speaker from just her speech. Specifically, the time-varying contextual non-linear mapping between audio stream and visual facial movements is realized by training a LSTM neural network on a large audio-visual data corpus. In this work, we extract a set of acoustic features from input audio, including Mel-scaled spectrogram, Mel frequency cepstral coefficients and chromagram that can effectively represent both contextual progression and emotional intensity of the speech. Output facial movements are characterized by 3D rotation and blending expression weights of a blendshape model, which can be used directly for animation. Thus, even though our model does not explicitly predict the affective states of the target speaker, her emotional manifestation is recreated via expression weights of the face model. Experiments on an evaluation dataset of different speakers across a wide range of affective states demonstrate promising results of our approach in real-time speech-driven facial animation.
机译:我们为实时面部动画引入了长短期记忆循环神经网络(LSTM-RNN)方法,该方法可以仅根据讲话者的语音自动估计其头部旋转和面部动作单元激活。具体来说,通过在大型视听数据语料库上训练LSTM神经网络,可以实现音频流和视觉面部运动之间随时间变化的上下文非线性映射。在这项工作中,我们从输入音频中提取了一组声学特征,包括梅尔标度频谱图,梅尔频率倒谱系数和色谱图,它们可以有效地代表语境的发展和语音的情感强度。输出面部运动的特征是3D旋转和blendshape模型的混合表达权重,可以直接用于动画。因此,即使我们的模型没有明确预测目标说话者的情感状态,但她的情感表现还是通过面部模型的表情权重来重新创建的。在广泛的情感状态下,对不同说话者的评估数据集进行的实验表明,我们的方法在实时语音驱动的面部动画中具有令人鼓舞的结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号