首页> 外文会议>IEEE Conference on Computer Vision and Pattern Recognition >Efficient Piecewise Training of Deep Structured Models for Semantic Segmentation
【24h】

Efficient Piecewise Training of Deep Structured Models for Semantic Segmentation

机译:深度分割的语义分段的有效分段训练

获取原文

摘要

Recent advances in semantic image segmentation have mostly been achieved by training deep convolutional neural networks (CNNs). We show how to improve semantic segmentation through the use of contextual information, specifically, we explore 'patch-patch' context between image regions, and 'patch-background' context. For learning from the patch-patch context, we formulate Conditional Random Fields (CRFs) with CNN-based pairwise potential functions to capture semantic correlations between neighboring patches. Efficient piecewise training of the proposed deep structured model is then applied to avoid repeated expensive CRF inference for back propagation. For capturing the patch-background context, we show that a network design with traditional multi-scale image input and sliding pyramid pooling is effective for improving performance. Our experimental results set new state-of-the-art performance on a number of popular semantic segmentation datasets, including NYUDv2, PASCAL VOC 2012, PASCAL-Context, and SIFT-flow. In particular, we achieve an intersection-overunion score of 78:0 on the challenging PASCAL VOC 2012 dataset.
机译:语义图像分割的最新进展主要是通过训练深度卷积神经网络(CNN)来实现的。我们展示了如何通过使用上下文信息来改善语义分割,特别是,我们探索了图像区域之间的“补丁-补丁”上下文和“补丁-背景”上下文。为了从补丁补丁上下文中学习,我们用基于CNN的成对潜在函数来公式化条件随机字段(CRF),以捕获相邻补丁之间的语义相关性。然后,对提出的深层结构模型进行有效的分段训练,以避免重复昂贵的CRF推论回传。为了捕获补丁背景,我们表明使用传统的多尺度图像输入和滑动金字塔池进行网络设计可以有效地提高性能。我们的实验结果为许多流行的语义分割数据集(包括NYUDv2,PASCAL VOC 2012,PASCAL-Context和SIFT-flow)设置了最新的性能。特别是,我们在具有挑战性的PASCAL VOC 2012数据集上获得了78:0的相交重叠分数。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号