首页> 外文会议>European conference on computer vision >Predicting Gaze in Egocentric Video by Learning Task-Dependent Attention Transition
【24h】

Predicting Gaze in Egocentric Video by Learning Task-Dependent Attention Transition

机译:通过学习任务相关的注意力转移来预测以自我为中心的视频中的凝视

获取原文

摘要

We present a new computational model for gaze prediction in egocentric videos by exploring patterns in temporal shift of gaze fixations (attention transition) that are dependent on egocentric manipulation tasks. Our assumption is that the high-level context of how a task is completed in a certain way has a strong influence on attention transition and should be modeled for gaze prediction in natural dynamic scenes. Specifically, we propose a hybrid model based on deep neural networks which integrates task-dependent attention transition with bottom-up saliency prediction. In particular, the task-dependent attention transition is learned with a recurrent neural network to exploit the temporal context of gaze fixations, e.g. looking at a cup after moving gaze away from a grasped bottle. Experiments on public egocentric activity datasets show that our model significantly outperforms state-of-the-art gaze prediction methods and is able to learn meaningful transition of human attention.
机译:通过探索依赖于以自我为中心的操纵任务的凝视注视的时间变化(注意力转移)的模式,我们提出了以自我为中心的视频中凝视预测的新计算模型。我们的假设是,以某种方式完成任务的高级上下文对注意力转移有很大影响,因此应该为自然动态场景中的凝视预测建模。具体来说,我们提出了一种基于深度神经网络的混合模型,该模型将任务依赖的注意力转移与自下而上的显着性预测相结合。特别地,通过递归神经网络学习依赖于任务的注意力转移,以利用凝视注视的时间上下文,例如视线凝视。将目光从紧紧抓住的瓶子上移开后,看着杯子。在以公众为中心的活动数据集上进行的实验表明,我们的模型明显优于最新的注视预测方法,并且能够学习有意义的人类注意力转移。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号