首页> 外文会议>2018 13th IEEE International Conference on Automatic Face amp; Gesture Recognition >Expressive Speech-Driven Lip Movements with Multitask Learning
【24h】

Expressive Speech-Driven Lip Movements with Multitask Learning

机译:具有多任务学习能力的语音驱动唇部动作

获取原文
获取原文并翻译 | 示例

摘要

The orofacial area conveys a range of information, including speech articulation and emotions. These two factors add constraints to the facial movements, creating non-trivial integrations and interplays. To generate more expressive and naturalistic movements for conversational agents (CAs) the relationship between these factors should be carefully modeled. Data-driven models are more appropriate for this task than rule-based systems. This paper provides two deep learning speech-driven structures to integrate speech articulation and emotional cues. The proposed approaches rely on multitask learning (MTL) strategies, where related secondary tasks are jointly solved when synthesizing orofacial movements. In particular, we evaluate emotion recognition and viseme recognition as secondary tasks. The approach creates shared representations that generate behaviors that not only are closer to the original orofacial movements, but also are perceived more natural than the results from single task learning.
机译:口腔区域传达了一系列信息,包括语音清晰度和情感。这两个因素增加了面部运动的约束,创造了非同寻常的融合和相互作用。为了使对话代理(CA)产生更具表现力和自然主义的运动,应该仔细模拟这些因素之间的关系。数据驱动模型比基于规则的系统更适合此任务。本文提供了两种深度学习语音驱动的结构,以整合语音清晰度和情感提示。所提出的方法依赖于多任务学习(MTL)策略,其中在合成口腔运动时共同解决相关的次要任务。尤其是,我们将情感识别和视位识别视为次要任务。该方法创建共享的表示,这些表示产生的行为不仅比原始的口腔运动更接近,而且比单个任务学习的结果更自然。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号