首页> 美国卫生研究院文献>Frontiers in Computational Neuroscience >Applying artificial vision models to human scene understanding
【2h】

Applying artificial vision models to human scene understanding

机译:将人工视觉模型应用于人类场景理解

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

How do we understand the complex patterns of neural responses that underlie scene understanding? Studies of the network of brain regions held to be scene-selective—the parahippocampal/lingual region (PPA), the retrosplenial complex (RSC), and the occipital place area (TOS)—have typically focused on single visual dimensions (e.g., size), rather than the high-dimensional feature space in which scenes are likely to be neurally represented. Here we leverage well-specified artificial vision systems to explicate a more complex understanding of how scenes are encoded in this functional network. We correlated similarity matrices within three different scene-spaces arising from: (1) BOLD activity in scene-selective brain regions; (2) behavioral measured judgments of visually-perceived scene similarity; and (3) several different computer vision models. These correlations revealed: (1) models that relied on mid- and high-level scene attributes showed the highest correlations with the patterns of neural activity within the scene-selective network; (2) NEIL and SUN—the models that best accounted for the patterns obtained from PPA and TOS—were different from the GIST model that best accounted for the pattern obtained from RSC; (3) The best performing models outperformed behaviorally-measured judgments of scene similarity in accounting for neural data. One computer vision method—NEIL (“Never-Ending-Image-Learner”), which incorporates visual features learned as statistical regularities across web-scale numbers of scenes—showed significant correlations with neural activity in all three scene-selective regions and was one of the two models best able to account for variance in the PPA and TOS. We suggest that these results are a promising first step in explicating more fine-grained models of neural scene understanding, including developing a clearer picture of the division of labor among the components of the functional scene-selective brain network.
机译:我们如何理解构成场景理解基础的神经反应的复杂模式?被认为是场景选择的大脑区域网络的研究(海马旁/舌侧区域(PPA),脾后复合体(RSC)和枕骨部位区域(TOS))通常集中于单个视觉维度(例如大小) ),而不是可能用神经表示场景的高维特征空间。在这里,我们利用规范化的人工视觉系统来阐明对在此功能网络中如何编码场景的更复杂的理解。我们将三个不同场景空间内的相似度矩阵相关联:(1)场景选择性大脑区域的BOLD活动; (2)行为测量的视觉感知场景相似性判断; (3)几种不同的计算机视觉模型。这些相关性揭示了:(1)依赖于中高层场景属性的模型显示出与场景选择网络中神经活动模式的最高相关性; (2)NEIL和SUN是最能说明从PPA和TOS获得的模式的模型,而GIST模型与最能说明从RSC获得的模式的模型不同。 (3)在计算神经数据方面,性能最好的模型优于对场景相似性进行行为测量的判断。一种计算机视觉方法-NEIL(“永无止境-图像学习器”)结合了视觉特征,这些视觉特征作为跨网络规模场景的统计规律而学习,在所有三个场景选择区域中均与神经活动具有显着相关性,并且是一种这两种模型中最能说明PPA和TOS差异的模型。我们认为,这些结果是阐明更精细的神经场景理解模型的有希望的第一步,包括对功能场景选择性大脑网络的各个组成部分之间的分工进行更清晰的描述。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号