首页> 外文OA文献 >Unsupervised learning of invariant representations with low sample complexity: the magic of sensory cortex or a new framework for machine learning?
【2h】

Unsupervised learning of invariant representations with low sample complexity: the magic of sensory cortex or a new framework for machine learning?

机译:无监督学习低样本复杂度的不变表示:感知皮层的神奇之处还是机器学习的新框架?

摘要

The present phase of Machine Learning is characterized by supervised learning algorithms relying on large sets of labeled examples (n → ∞). The next phase is likely to focus on algorithms capable of learning from very few labeled examples (n → ∞), like humans seem able to do. We propose an approach to this problem and describe the underlying theory, based on the unsupervised, automatic learning of a "good" representation for supervised learning, characterized by small sample complexity (n). We consider the case of visual object recognition though the theory applies to otherdomains. The starting point is the conjecture, proved in specific cases, that image representations which are invariant to translations, scaling and other transformations can considerably reduce the sample complexity of learning. We prove that an invariant and unique (discriminative) signature can be computed for each image patch, I, in terms of empirical distributionsof the dot-products between I and a set of templates stored during unsupervised learning. A module performing filtering and pooling, like the simple and complex cells described by Hubel and Wiesel, can compute such estimates. Hierarchical architectures consisting of this basic Hubel-Wiesel moduli inherit its properties of invariance, stability, and discriminabilitywhile capturing the compositional organization of the visual world in terms of wholes and parts. The theory extends existing deep learning convolutional architectures for image and speech recognition. It also suggests that the main computational goal of the ventral stream of visual cortex is to provide a hierarchical representation of new objects/images which is invariantto transformations, stable, and discriminative for recognition|and that this representation may be continuously learned in an unsupervised way during development and visual experience.
机译:机器学习的当前阶段的特征在于,监督学习算法依赖于大量带标签的示例(n→∞)。下一阶段可能将重点放在能够从很少的标记示例(n→∞)中学习的算法上,就像人类似乎能够做到的那样。我们提出了一个解决此问题的方法,并基于无监督,自动学习“良好”表示的有监督学习的“良好”表示形式,描述了基础理论,其特征是样本复杂度较小(n)。尽管该理论适用于其他领域,但我们考虑了视觉对象识别的情况。起点是在特定情况下证明的猜想,即对于平移,缩放和其他变换不变的图像表示可以显着降低学习的样本复杂性。我们证明,根据I与非监督学习过程中存储的一组模板之间的点积的经验分布,可以为每个图像补丁I计算出不变且唯一的(区分性)签名。像Hubel和Wiesel所描述的简单和复杂单元格一样,执行过滤和合并的模块可以计算此类估计。由基本的Hubel-Wiesel模量组成的分层体系结构继承了其不变性,稳定性和可区分性的属性,同时从整体和部分的角度捕获了视觉世界的组成组织。该理论扩展了用于图像和语音识别的现有深度学习卷积架构。还建议视觉皮层腹侧流的主要计算目标是提供新对象/图像的层次表示形式,该新对象/图像对于变换而言是不变的,稳定的和可辨别的,并且该表示形式可以以无监督的方式不断学习。在开发和视觉体验中。

著录项

  • 作者

  • 作者单位
  • 年度 2014
  • 总页数
  • 原文格式 PDF
  • 正文语种 en_US
  • 中图分类

相似文献

  • 外文文献
  • 中文文献
  • 专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号