首页> 外文会议>European Conference on Computer Vision >Forecasting Human-Object Interaction: Joint Prediction of Motor Attention and Actions in First Person Video
【24h】

Forecasting Human-Object Interaction: Joint Prediction of Motor Attention and Actions in First Person Video

机译:预测人对象互动:第一人体视频中电机关注的联合预测

获取原文

摘要

We address the challenging task of anticipating human-object interaction in first person videos. Most existing methods either ignore how the camera wearer interacts with objects, or simply considers body motion as a separate modality. In contrast, we observe that the intentional hand movement reveals critical information about the future activity. Motivated by this observation, we adopt intentional hand movement as a feature representation, and propose a novel deep network that jointly models and predicts the egocentric hand motion, interaction hotspots and future action. Specifically, we consider the future hand motion as the motor attention, and model this attention using probabilistic variables in our deep model. The predicted motor attention is further used to select the discriminative spatial-temporal visual features for predicting actions and interaction hotspots. We present extensive experiments demonstrating the benefit of the proposed joint model. Importantly, our model produces new state-of-the-art results for action anticipation on both EGTEA Gaze+ and the EPIC-Kitchens datasets.
机译:我们解决了在第一人称视频中预测人对象互动的具有挑战性的任务。大多数现有方法既忽略相机佩戴者如何与对象交互,或者只是将身体运动视为单独的模态。相比之下,我们观察到故意的手势揭示了关于未来活动的关键信息。通过这种观察,我们采用故意手工作为特征表示,并提出了一种新的深度网络,共同模型和预测Epentric手动运动,交互热点和未来动作。具体而言,我们考虑未来的手动运动作为电机的注意力,并在我们深层模型中使用概率变量来模拟这种注意。预测的电机注意力还用于选择用于预测动作和交互热点的判别空间 - 时间视觉特征。我们呈现出广泛的实验,证明了提出的联合模型的利益。重要的是,我们的模型为Egtea Gaze +和史诗厨房数据集产生了新的最先进结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号