首页> 外文会议>Annual Conference on Neural Information Processing Systems >Congruence between model and human attention reveals unique signatures of critical visual events
【24h】

Congruence between model and human attention reveals unique signatures of critical visual events

机译:模型与人类注意力之间的一致性揭示了关键视觉事件的独特特征

获取原文

摘要

Current computational models of bottom-up and top-down components of attention are predictive of eye movements across a range of stimuli and of simple, fixed visual tasks (such as visual search for a target among distractors). However, to date there exists no computational framework which can reliably mimic human gaze behavior in more complex environments and tasks, such as driving a vehicle through traffic. Here, we develop a hybrid computational/behavioral framework, combining simple models for bottom-up salience and top-down relevance, and looking for changes in the predictive power of these components at different critical event times during 4.7 hours (500,000 video frames) of observers playing car racing and flight combat video games. This approach is motivated by our observation mat the predictive strengths of the salience and relevance models exhibit reliable temporal signatures during critical event windows in the task sequence-for example, when the game player directly engages an enemy plane in a flight combat game, the predictive strength of the salience model increases significantly, while that of the relevance model decreases significantly. Our new framework combines these temporal signatures to implement several event detectors. Critically, we find that an event detector based on fused behavioral and stimulus information (in the form of the model's predictive strength) is much stronger than detectors based on behavioral information alone (eye position) or image information alone (model prediction maps). This approach to event detection, based on eye tracking combined with computational models applied to the visual input, may have useful applications as a less-invasive alternative to other event detection approaches based on neural signatures derived from EEG or fMRI recordings.
机译:当前,注意力的自上而下和自上而下的计算模型可以预测眼睛在一系列刺激下的运动以及简单,固定的视觉任务(例如,对干扰对象的视觉搜索)。但是,迄今为止,还没有一种计算框架能够可靠地模仿更复杂的环境和任务(例如通过交通驾驶车辆)中的人的凝视行为。在这里,我们开发了一个混合的计算/行为框架,结合了自下而上的显着性和自上而下的相关性的简单模型,并在4.7小时(500,000个视频帧)的不同关键事件时间寻找这些组件的预测能力的变化观察者玩赛车和飞行战斗视频游戏。这种方法是由我们的观察所激发的,显着性和相关性模型的预测强度在任务序列中的关键事件窗口期间显示出可靠的时间特征-例如,当玩家在飞行战斗游戏中直接与敌机交战时,预测性显着性模型的强度显着增加,而相关性模型的强度显着降低。我们的新框架结合了这些时间签名,以实现多个事件检测器。至关重要的是,我们发现基于行为和刺激信息融合(以模型的预测强度的形式)的事件检测器比仅基于行为信息(眼位)或仅基于图像信息(模型预测图)的检测器强得多。这种基于眼睛跟踪与应用于视觉输入的计算模型相结合的事件检测方法,作为基于基于EEG或fMRI记录的神经签名的其他事件检测方法的侵入性较小的替代方法,可能具有有用的应用。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号