首页> 美国卫生研究院文献>Frontiers in Computational Neuroscience >Finding and recognizing objects in natural scenes: complementary computations in the dorsal and ventral visual systems
【2h】

Finding and recognizing objects in natural scenes: complementary computations in the dorsal and ventral visual systems

机译:在自然场景中寻找和识别物体:背侧和腹侧视觉系统中的互补计算

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Searching for and recognizing objects in complex natural scenes is implemented by multiple saccades until the eyes reach within the reduced receptive field sizes of inferior temporal cortex (IT) neurons. We analyze and model how the dorsal and ventral visual streams both contribute to this. Saliency detection in the dorsal visual system including area LIP is modeled by graph-based visual saliency, and allows the eyes to fixate potential objects within several degrees. Visual information at the fixated location subtending approximately 9° corresponding to the receptive fields of IT neurons is then passed through a four layer hierarchical model of the ventral cortical visual system, VisNet. We show that VisNet can be trained using a synaptic modification rule with a short-term memory trace of recent neuronal activity to capture both the required view and translation invariances to allow in the model approximately 90% correct object recognition for 4 objects shown in any view across a range of 135° anywhere in a scene. The model was able to generalize correctly within the four trained views and the 25 trained translations. This approach analyses the principles by which complementary computations in the dorsal and ventral visual cortical streams enable objects to be located and recognized in complex natural scenes.
机译:搜索和识别复杂自然场景中的对象是通过多次扫视实现的,直到眼睛到达下颞叶皮层(IT)神经元的减小的接受视野范围之内。我们分析并模拟了背侧和腹侧视觉流如何对此做出贡献。包括区域LIP的背侧视觉系统中的显着性检测是通过基于图形的视觉显着性建模的,并允许眼睛将潜在物体固定在几个角度之内。然后,通过与腹侧皮层视觉系统VisNet的四层分层模型传递固定位置的可视信息,该位置大约对应于IT神经元的感受野大约9°。我们显示,可以使用具有最近神经元活动的短期记忆痕迹的突触修改规则来训练VisNet,以捕获所需的视图和翻译不变性,从而在模型中允许对任何视图中显示的4个对象进行大约90%的正确对象识别在场景中任何地方的135°范围内。该模型能够在4个经过训练的视图和25个经过训练的翻译中正确地进行概括。这种方法分析了背侧和腹侧视觉皮层流中的互补计算使物体能够在复杂的自然场景中定位和识别的原理。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号