首页> 外文会议>European Conference on Computer Vision >Real-Time Facial Segmentation and Performance Capture from RGB Input
【24h】

Real-Time Facial Segmentation and Performance Capture from RGB Input

机译:RGB输入的实时面部分段和性能捕获

获取原文

摘要

We introduce the concept of unconstrained real-time 3D facial performance capture through explicit semantic segmentation in the RGB input. To ensure robustness, cutting edge supervised learning approaches rely on large training datasets of face images captured in the wild. While impressive tracking quality has been demonstrated for faces that are largely visible, any occlusion due to hair, accessories, or hand-to-face gestures would result in significant visual artifacts and loss of tracking accuracy. The modeling of occlusions has been mostly avoided due to its immense space of appearance variability. To address this curse of high dimensionality, we perform tracking in unconstrained images assuming non-face regions can be fully masked out. Along with recent breakthroughs in deep learning, we demonstrate that pixel-level facial segmentation is possible in real-time by repurposing convolutional neural networks designed originally for general semantic segmentation. We develop an efficient architecture based on a two-stream deconvolution network with complementary characteristics, and introduce carefully designed training samples and data augmentation strategies for improved segmentation accuracy and robustness. We adopt a state-of-the-art regression-based facial tracking framework with segmented face images as training, and demonstrate accurate and uninterrupted facial performance capture in the presence of extreme occlusion and even side views. Furthermore, the resulting segmentation can be directly used to composite partial 3D face models on the input images and enable seamless facial manipulation tasks, such as virtual make-up or face replacement.
机译:我们通过RGB输入中的显式语义分段介绍无约束的实时3D面部性能捕获的概念。为确保鲁棒性,切削刃监督学习方法依赖于在野外捕获的面部图像的大型训练数据集。虽然对基本上可见的面部展示了令人印象深刻的跟踪质量,但由于头发,配件或手动手势而导致的任何遮挡都会导致显着的视觉伪影和跟踪精度的损失。由于其巨大的外观变异性,因此,闭塞的建模主要是避免。为了解决高维度的诅咒,我们在假设非面积区域可以完全掩盖非面积区域的无约会图像中执行跟踪。随着近期深度学习的突破,我们证明了像素级面部分割是可以实时进行的,通过重新调整最初用于一般语义分割的卷积神经网络。我们基于具有互补特性的双流碎片卷积网络的高效架构,并介绍精心设计的培训样本和数据增强策略,以提高分割精度和鲁棒性。我们采用基于最先进的回归的面部跟踪框架,作为培训分段面部图像,并在极端遮挡甚至侧视图存在下展示准确和不间断的面部性能捕获。此外,所得到的分割可以直接用于复合输入图像上的部分3D面部模型,并使无缝面部操纵任务,例如虚拟化妆或面部更换。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号