首页> 美国卫生研究院文献>PLoS Clinical Trials >Ultra-rapid object categorization in real-world scenes with top-down manipulations
【2h】

Ultra-rapid object categorization in real-world scenes with top-down manipulations

机译:通过自上而下的操作在真实场景中实现超快速的对象分类

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Humans are able to achieve visual object recognition rapidly and effortlessly. Object categorization is commonly believed to be achieved by interaction between bottom-up and top-down cognitive processing. In the ultra-rapid categorization scenario where the stimuli appear briefly and response time is limited, it is assumed that a first sweep of feedforward information is sufficient to discriminate whether or not an object is present in a scene. However, whether and how feedback/top-down processing is involved in such a brief duration remains an open question. To this end, here, we would like to examine how different top-down manipulations, such as category level, category type and real-world size, interact in ultra-rapid categorization. We have constructed a dataset comprising real-world scene images with a built-in measurement of target object display size. Based on this set of images, we have measured ultra-rapid object categorization performance by human subjects. Standard feedforward computational models representing scene features and a state-of-the-art object detection model were employed for auxiliary investigation. The results showed the influences from 1) animacy (animal, vehicle, food), 2) level of abstraction (people, sport), and 3) real-world size (four target size levels) on ultra-rapid categorization processes. This had an impact to support the involvement of top-down processing when rapidly categorizing certain objects, such as sport at a fine grained level. Our work on human vs. model comparisons also shed light on possible collaboration and integration of the two that may be of interest to both experimental and computational vision researches. All the collected images and behavioral data as well as code and models are publicly available at .
机译:人类能够快速,轻松地实现视觉目标识别。通常认为,对象分类是通过自下而上和自上而下的认知处理之间的交互来实现的。在短暂出现刺激且响应时间有限的超快速分类场景中,假定前馈信息的第一次扫描足以区分场景中是否存在对象。但是,在这么短的时间内是否涉及反馈方式和自上而下的处理仍是一个悬而未决的问题。为此,在这里,我们要研究不同的自上而下的操作(例如类别级别,类别类型和实际大小)如何在超快速分类中进行交互。我们构建了一个包含真实场景图像的数据集,并内置了对目标物体显示尺寸的测量。基于这组图像,我们已经测量了人类对象的超快速物体分类性能。代表场景特征的标准前馈计算模型和最新的物体检测模型被用于辅助研究。结果表明:1)动画(动物,交通工具,食物),2)抽象级别(人,运动)和3)实际大小(四个目标大小级别)对超快速分类过程的影响。当快速对某些对象(例如运动)进行细分类时,这有助于支持自上而下的处理。我们在人与模型比较方面的工作还阐明了两者之间可能的协作和整合,这可能对实验和计算视觉研究都非常感兴趣。所有收集的图像和行为数据以及代码和模型都可以在上公开获得。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号