首页> 外文会议>Conference on wavelets and sparsity XV >Rotation-covariant visual concept detection using steerable Riesz wavelets and bags of visual words
【24h】

Rotation-covariant visual concept detection using steerable Riesz wavelets and bags of visual words

机译:旋转 - 协调性视觉概念检测使用可转向riesz小波和视觉单词袋

获取原文

摘要

Distinct texture classes are often sharing several visual concepts. Texture instances from different classes are sharing regions in the feature hyperspace, which results in ill-defined classification configurations. In this work, we detect rotation-covariant visual concepts using steerable Riesz wavelets and bags of visual words. In a first step, K-means clustering is used to detect visual concepts in the hyperspace of the energies of steerable Riesz wavelets. The coordinates of the clusters are used to construct templates from linear combinations of the Riesz components that are corresponding to visual concepts. The visualization of these templates allows verifying the relevance of the concepts modeled. Then, the local orientations of each template are optimized to maximize their response, which is carried out analytically and can still be expressed as a linear combination of the initial steerable Riesz templates. The texture classes are learned in the feature space composed of the concatenation of the maximum responses of each visual concept using support vector machines. An experimental evaluation using the Outex_TC_00010 test suite allowed a classification accuracy of 97.5%, which demonstrates the feasibility of the proposed approach. An optimal number K = 20 of clusters is required to model the visual concepts, which was found to be fewer than the number of classes. This shows that higher-level classes are sharing low-level visual concepts. The importance of rotation-covariant visual concept modeling is highlighted by allowing an absolute gain of more than 30% in accuracy. The visual concepts are modeling the local organization of directions at various scales, which is in accordance with the bottom-up visual information processing sequence of the primal sketch in Marr's theory on vision.
机译:独特的纹理课程通常共享几种视觉概念。来自不同类的纹理实例是在特征超空间中共享区域,从而导致定义了定义的分类配置。在这项工作中,我们使用可操纵的riesz小波和袋子的视觉词来检测旋转 - 协助视觉概念。在第一步中,K-means聚类用于检测可操纵的Riesz小波能量的高静脉空间中的视觉概念。群集的坐标用于构造来自对应于视觉概念的RIESZ组件的线性组合构造模板。这些模板的可视化允许验证所建模概念的相关性。然后,优化每个模板的本地取向以最大化其响应,其被分析执行,并且仍然可以表示为初始可转向riesz模板的线性组合。在使用支持向量机器的每个视觉概念的最大响应组成的特征空间中学习纹理类。使用Outex_TC_00010测试套件的实验评估允许分类精度为97.5%,这证明了所提出的方法的可行性。为模拟视觉概念需要最佳数字k = 20个集群,发现该概念被发现少于类的数量。这表明更高级别的类是共享低级视觉概念。通过精度允许绝对增益超过30%,突出了旋转协方面的视觉概念建模的重要性。视觉概念正在以各种尺度在各种尺度上建模局部方向组织,这符合Marr在Marr的视觉理论中的原始素描的自下而上的视觉信息处理序列。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号