...
首页> 外文期刊>ACM Transactions on Graphics >High-Fidelity Facial and Speech Animation for VR HMDs
【24h】

High-Fidelity Facial and Speech Animation for VR HMDs

机译:VR HMD的高保真面部和语音动画

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

Significant challenges currently prohibit expressive interaction in virtualrnreality (VR). Occlusions introduced by head-mounted displaysrn(HMDs) make existing facial tracking techniques intractable, andrneven state-of-the-art techniques used for real-time facial tracking inrnunconstrained environments fail to capture subtle details of the user’srnfacial expressions that are essential for compelling speech animation.rnWe introduce a novel system for HMD users to control a digitalrnavatar in real-time while producing plausible speech animation andrnemotional expressions. Using a monocular camera attached to anrnHMD, we record multiple subjects performing various facial expressionsrnand speaking several phonetically-balanced sentences. Thesernimages are used with artist-generated animation data correspondingrnto these sequences to train a convolutional neural network (CNN)rnto regress images of a user’s mouth region to the parameters thatrncontrol a digital avatar. To make training this system more tractable,rnwe use audio-based alignment techniques to map images of multiplernusers making the same utterance to the corresponding animationrnparameters. We demonstrate that this approach is also feasible forrntracking the expressions around the user’s eye region with an internalrninfrared (IR) camera, thereby enabling full facial tracking. Thisrnsystem requires no user-specific calibration, uses easily obtainablernconsumer hardware, and produces high-quality animations of speechrnand emotional expressions. Finally, we demonstrate the quality ofrnour system on a variety of subjects and evaluate its performancernagainst state-of-the-art real-time facial tracking techniques.
机译:当前,重大挑战禁止虚拟现实(VR)中的表达交互。头戴式显示器(HMD)引入的遮挡使现有的面部跟踪技术变得难以处理,并且即使在不受约束的环境下用于实时面部跟踪的最新技术也无法捕捉到用户面部表情的细微细节,而这些细节对于引人入胜的语音动画。我们为HMD用户引入了一种新颖的系统,可以实时控制数字化身,同时产生合理的语音动画和情感表达。使用连接到anrnHMD的单眼相机,我们记录了执行各种面部表情的多个对象,并讲了几个语音平衡的句子。这些图像与与这些序列相对应的艺术家生成的动画数据一起使用,以训练卷积神经网络(CNN)以将用户口部区域的图像回归到控制数字化身的参数。为了使该系统的训练更容易处理,我们使用基于音频的对齐技术将发声相同的多个用户的图像映射到相应的动画参数。我们证明了这种方法对于使用内置红外(IR)摄像机跟踪用户眼睛区域周围的表情也是可行的,从而可以进行全面的面部跟踪。该系统不需要特定于用户的校准,使用易于获得的消费硬件,并可以生成高质量的语音和情感表达动画。最后,我们在各种主题上演示了nour系统的质量,并针对最先进的实时面部跟踪技术评估了其性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号