首页> 外文会议>European conference on computer vision >Multiple-Gaze Geometry: Inferring Novel 3D Locations from Gazes Observed in Monocular Video
【24h】

Multiple-Gaze Geometry: Inferring Novel 3D Locations from Gazes Observed in Monocular Video

机译:多凝视几何:从单眼视频中观察到的凝视推断出新颖的3D位置

获取原文

摘要

We develop using person gaze direction for scene understanding. In particular, we use intersecting gazes to learn 3D locations that people tend to look at, which is analogous to having multiple camera views. The 3D locations that we discover need not be visible to the camera. Conversely, knowing 3D locations of scene elements that draw visual attention, such as other people in the scene, can help infer gaze direction. We provide a Bayesian generative model for the temporal scene that captures the joint probability of camera parameters, locations of people, their gaze, what they are looking at, and locations of visual attention. Both the number of people in the scene and the number of extra objects that draw attention are unknown and need to be inferred. To execute this joint inference we use a probabilistic data association approach that enables principled comparison of model hypotheses. We use MCMC for inference over the discrete correspondence variables, and approximate the marginalization over continuous parameters using the Metropolis-Laplace approximation, using Hamiltonian (Hybrid) Monte Carlo for maximization. As existing data sets do not provide the 3D locations of what people are looking at, we contribute a small data set that does. On this data set, we infer what people are looking at with 59% precision compared with 13% for a baseline approach, and where those objects are within about 0.58 m.
机译:为场景理解使用人的注视方向进行开发。特别是,我们使用相交的凝视来学习人们倾向于观看的3D位置,这类似于具有多个相机视图。我们发现的3D位置不需要在相机中可见。相反,了解引起视觉注意的场景元素(例如场景中的其他人)的3D位置可以帮助推断注视方向。我们为时态场景提供了贝叶斯生成模型,该模型捕获了摄像机参数,人物位置,他们的注视,他们正在看的东西以及视觉注意力的位置的联合概率。场景中的人数和引起注意的其他物体的数量都是未知的,需要进行推断。为了执行此联合推断,我们使用概率数据关联方法,该方法可以对模型假设进行原则上的比较。我们使用MCMC推断离散的对应变量,并使用Metropolis-Laplace逼近近似连续参数的边际化,使用Hamiltonian(Hybrid)Monte Carlo求最大化。由于现有数据集无法提供人们正在查看的3D位置,因此我们贡献了一个小型数据集。在此数据集上,我们推断人们所看到的东西的准确度为59%,而基线方法的准确度为13%,并且这些物体在大约0.58 m的范围内。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号