首页> 外文会议>IEEE/CVF Conference on Computer Vision and Pattern Recognition >Actor and Observer: Joint Modeling of First and Third-Person Videos
【24h】

Actor and Observer: Joint Modeling of First and Third-Person Videos

机译:演员和观察员:第一人称和第三人称视频的联合建模

获取原文

摘要

Several theories in cognitive neuroscience suggest that when people interact with the world, or simulate interactions, they do so from a first-person egocentric perspective, and seamlessly transfer knowledge between third-person (observer) and first-person (actor). Despite this, learning such models for human action recognition has not been achievable due to the lack of data. This paper takes a step in this direction, with the introduction of Charades-Ego, a large-scale dataset of paired first-person and third-person videos, involving 112 people, with 4000 paired videos. This enables learning the link between the two, actor and observer perspectives. Thereby, we address one of the biggest bottlenecks facing egocentric vision research, providing a link from first-person to the abundant third-person data on the web. We use this data to learn a joint representation of first and third-person videos, with only weak supervision, and show its effectiveness for transferring knowledge from the third-person to the first-person domain.
机译:认知神经科学中的几种理论表明,当人们与世界互动或模拟互动时,他们是从第一人称自我中心角度进行的,并在第三人称(观察者)和第一人称(演员)之间无缝地传递知识。尽管如此,由于缺乏数据,仍无法实现这种用于人类动作识别的模型。本文朝着这个方向迈出了一步,引入了Charades-Ego,这是一个由第一人称和第三人称视频配对的大规模数据集,涉及112人,有4000个配对视频。这使得能够了解演员和观察者视角之间的联系。因此,我们解决了以自我为中心的视觉研究面临的最大瓶颈之一,它提供了从第一人称视角到网络上丰富的第三人称数据的链接。我们使用此数据来学习第一人称视频和第三人称视频的联合表示,而仅需很少的监督,并显示其将知识从第三人称转移到第一人称领域的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号