首页> 外文期刊>Journal of vision >Measuring the salience of an object in a scene
【24h】

Measuring the salience of an object in a scene

机译:测量场景中物体的显着性

获取原文
           

摘要

Over the past 15 years work on visual salience has been restricted to models of low-level, bottom-up salience that give an estimate of the salience for every pixel in an image. This study concerns the question of how to measure the salience of objects. More precisely, given an image and a list of areas of interest (AOIs), can we assign salience scores to the AOIs that reflect their visual prominence? Treating salience as a per-object feature allows us to incorporate a notion of salience into higher-level, cognitive models. There is increasing evidence that fixations locations are best explained at an object level [Einhauser et al 2008, JoV; Nuthmann & Henderson 2010, JoV] and an object-level notion of visual salience can be easily incorporated with other object features representing semantics [Hwang et al 2011, VisRes; Greene 2013, FrontiersPsych] and task relevance]. Extracting scores for AOIs from the saliency maps that are output by existing models is a non-trivial task. Using simple psychophysical (1/f-noise) stimuli, we demonstrate that simple methods for assigning salience score to AOIs (such as taking the maxima, mean, or sum of the relevant pixels in the salience map) produce unintuitive results, such as predicting that larger objects are less salient. We also evaluate object salience models over a range of tasks and compare to empirical data. Beyond predicting the number of fixations to different objects in a scene, we also estimate the difficulty of visual search trials; and incorporate visual salience into language production tasks. We present a simple object-based salience model (based on comparing the likelihood of an AOI given the rest of the image to the likelihood of a typical patch of the same area) that gives intuitive results for the 1/f-noise stimuli and performs as well as existing methods on empirical datasets.
机译:在过去的15年中,有关视觉显着性的工作仅限于低级,自下而上的显着性模型,这些模型可以估计图像中每个像素的显着性。这项研究涉及如何测量物体的显着性的问题。更准确地说,给定一张图像和感兴趣的区域(AOI)列表,我们是否可以为AOI分配显着性分数以反映其视觉突出程度?将显着性视为每个对象的特征使我们能够将显着性概念整合到更高级别的认知模型中。越来越多的证据表明,固定物的位置最好在物体水平上得到解释[Einhauser et al 2008,JoV; Nuthmann&Henderson 2010,JoV]和视觉显着性的对象级概念可以很容易地与其他表示语义的对象特征结合在一起[Hwang et al 2011,VisRes; [Greene 2013,FrontiersPsych]和任务相关性]。从现有模型输出的显着性图中提取AOI分数是一项艰巨的任务。使用简单的心理生理(1 / f噪声)刺激,我们证明了将显着性分数分配给AOI的简单方法(例如,获取显着性图中相关像素的最大值,平均值或总和)会产生不直观的结果,例如预测较大的物体不那么突出。我们还评估了一系列任务中的对象显着性模型,并与经验数据进行了比较。除了预测场景中不同对象的注视次数之外,我们还估计了视觉搜索试验的难度;并将视觉显着性纳入语言制作任务。我们提出了一个简单的基于对象的显着性模型(基于将给定图像其余部分的AOI的可能性与相同区域的典型斑块的可能性进行比较),从而为1 / f噪声刺激提供直观的结果并执行以及经验数据集上的现有方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号