首页> 外文期刊>IEEE Transactions on Image Processing >Ontology-Based Semantic Image Segmentation Using Mixture Models and Multiple CRFs
【24h】

Ontology-Based Semantic Image Segmentation Using Mixture Models and Multiple CRFs

机译:使用混合模型和多个CRF的基于本体的语义图像分割

获取原文
获取原文并翻译 | 示例
       

摘要

Semantic image segmentation is a fundamental yet challenging problem, which can be viewed as an extension of the conventional object detection with close relation to image segmentation and classification. It aims to partition images into non-overlapping regions that are assigned predefined semantic labels. Most of the existing approaches utilize and integrate low-level local features and high-level contextual cues, which are fed into an inference framework such as, the conditional random field (CRF). However, the lack of meaning in the primitives (i.e., pixels or superpixels) and the cues provides low discriminatory capabilities, since they are rarely object-consistent. Moreover, blind combinations of heterogeneous features and contextual cues exploitation through limited neighborhood relations in the CRFs tend to degrade the labeling performance. This paper proposes an ontology-based semantic image segmentation (OBSIS) approach that jointly models image segmentation and object detection. In particular, a Dirichlet process mixture model transforms the low-level visual space into an intermediate semantic space, which drastically reduces the feature dimensionality. These features are then individually weighed and independently learned within the context, using multiple CRFs. The segmentation of images into object parts is hence reduced to a classification task, where object inference is passed to an ontology model. This model resembles the way by which humans understand the images through the combination of different cues, context models, and rule-based learning of the ontologies. Experimental evaluations using the MSRC-21 and PASCAL VOC’2010 data sets show promising results.
机译:语义图像分割是一个基本而又具有挑战性的问题,可以看作是与图像分割和分类密切相关的常规对象检测的扩展。它旨在将图像划分为不重叠的区域,这些区域被分配了预定义的语义标签。大多数现有方法利用并集成了低级局部特征和高级上下文线索,这些低级局部特征和高级上下文线索被馈送到诸如条件随机字段(CRF)之类的推理框架中。但是,基元(即像素或超像素)和提示中缺乏意义,因为它们很少与对象保持一致,因此提供了较低的区分能力。而且,通过CRF中有限的邻域关系,异质特征和上下文线索开发的盲目组合往往会降低标记性能。本文提出了一种基于本体的语义图像分割(OBSIS)方法,将图像分割和目标检测联合建模。尤其是,狄利克雷(Dirichlet)过程混合模型将低级视觉空间转换为中间语义空间,从而大大降低了特征维数。然后,使用多个CRF在上下文中分别权衡并独立学习这些功能。因此,将图像分割成对象部分的工作减少到了分类任务,在该任务中,对象推断被传递给本体模型。该模型类似于人类通过不同线索,上下文模型和基于规则的本体学习的组合来理解图像的方式。使用MSRC-21和PASCAL VOC'2010数据集进行的实验评估显示出可喜的结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号