首页> 外文期刊>Multimedia Tools and Applications >Head and facial gestures synthesis using PAD model for an expressive talking avatar
【24h】

Head and facial gestures synthesis using PAD model for an expressive talking avatar

机译:使用PAD模型进行表情手势说话的头部和面部手势合成

获取原文
获取原文并翻译 | 示例

摘要

This paper proposes to synthesize expressive head and facial gestures on talking avatar using the three dimensional pleasure-displeasure, arousal-nonarousal and dominance-submissiveness (PAD) descriptors of semantic expressivity. The PAD model is adopted to bridge the gap between text semantics and visual motion features with three dimensions of pleasure-displeasure, arousal-nonarousal, and dominance-submissiveness. Based on the correlation analysis between PAD annotations and motion patterns derived from the head and facial motion database, we propose to build an explicit mapping from PAD descriptors to facial animation parameters with linear regression and neural networks for head motion and facial expression respectively. A PAD-driven talking avatar in text-to-visual-speech system is implemented by generating expressive head motions at the prosodic word level based on the (P, A) descriptors of lexical appraisal, and facial expressions at the sentence level according to the PAD descriptors of emotional information. A series of PAD reverse evaluation and comparative perceptual experiments shows that the head and facial gestures synthesized based on PAD model can significantly enhance the visual expressivity of talking avatar.
机译:本文提出使用语义表达的三维愉悦-不愉快,唤醒-非沉着和支配-顺从性(PAD)描述符在说话的化身上合成头部和面部手势。 PAD模型被用来弥合文本语义和视觉动作特征之间的差距,具有愉悦性,愉悦性,非情感性和支配性三个维度。基于从头部和面部运动数据库导出的PAD注释与运动模式之间的相关性分析,我们建议使用线性回归和神经网络分别针对头部运动和面部表情建立从PAD描述符到面部动画参数的显式映射。文本到视觉语音系统中的PAD驱动的说话化身是通过根据词汇评估的(P,A)描述符在韵律单词级别生成表达性头部动作以及根据句子级别在表情级别生成面部表情来实现的PAD情感信息描述符。一系列的PAD反向评估和对比知觉实验表明,基于PAD模型合成的头部和面部手势可以显着增强说话头像的视觉表现力。

著录项

  • 来源
    《Multimedia Tools and Applications》 |2014年第1期|439-461|共23页
  • 作者单位

    Key Laboratory of Pervasive Computing, Ministry of Education China, and Department of Computer Science and Technology, Tsinghua University, Beijing 100084, China;

    Department of Systems Engineering and Engineering Management, The Chinese University of Hong Kong (CUHK), Hong Kong, China,Tsinghua-CUHK Joint Research Center for Media Sciences, Technologies and Systems, Graduate School at Shenzhen, Tsinghua University, Shenzhen 518055, China;

    Key Laboratory of Pervasive Computing, Ministry of Education China, and Department of Computer Science and Technology, Tsinghua University, Beijing 100084, China;

    Department of Systems Engineering and Engineering Management, The Chinese University of Hong Kong (CUHK), Hong Kong, China,Tsinghua-CUHK Joint Research Center for Media Sciences, Technologies and Systems, Graduate School at Shenzhen, Tsinghua University, Shenzhen 518055, China;

    Key Laboratory of Pervasive Computing, Ministry of Education China, and Department of Computer Science and Technology, Tsinghua University, Beijing 100084, China;

  • 收录信息
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Text-to-visual-speech; Head motion; Facial expression; Talking avatar;

    机译:文字到视觉语音;头部运动;表情;说话头像;

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号