...
首页> 外文期刊>IEEE transactions on multimedia >Deep Co-Saliency Detection via Stacked Autoencoder-Enabled Fusion and Self-Trained CNNs
【24h】

Deep Co-Saliency Detection via Stacked Autoencoder-Enabled Fusion and Self-Trained CNNs

机译:通过堆叠的AutoEncoder的融合和自培训CNNS的深度合理检测

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

Image co-saliency detection via fusion-based or learning-based methods faces cross-cutting issues. Fusion-based methods often combine saliency proposals using a majority voting rule. Their performance hence highly depends on the quality and coherence of individual proposals. Learning-based methods typically require ground-truth annotations for training, which are not available for co-saliency detection. In this work, we present a two-stage approach to address these issues jointly. At the first stage, an unsupervised deep learning model with stacked autoencoder (SAE) is proposed to evaluate the quality of saliency proposals. It employs latent representations for image foregrounds, and auto-encodes foreground consistency and foreground-background distinctiveness in a discriminative way. The resultant model, SAE-enabled fusion (SAEF), can combine multiple saliency proposals to yield a more reliable saliency map. At the second stage, motivated by the fact that fusion often leads to over-smoothed saliency maps, we develop self-trained convolutional neural networks (STCNN) to alleviate this negative effect. STCNN takes the saliency maps produced by SAEF as inputs. It propagates information from regions of high confidence to those of low confidence. During propagation, feature representations are distilled, resulting in sharper and better co-saliency maps. Our approach is comprehensively evaluated on three benchmarks, including MSRC, iCoseg, and Cosal2015, and performs favorably against the state-of-the-arts. In addition, we demonstrate that our method can be applied to object co-segmentation and object co-localization, achieving the state-of-the-art performance in both applications.
机译:通过基于融合的或基于学习的方法的图像共同显着性检测面临跨切割问题。基于融合的方法通常使用大多数投票规则来组合显着性建议。因此,他们的表现非常取决于个别提案的质量和一致性。基于学习的方法通常需要训练的地面实际注释,这是不可用于共同显着性检测的训练。在这项工作中,我们提出了一种两级方法,共同解决这些问题。在第一阶段,提出了一种无监督的深度学习模型与堆叠的AutoEncoder(SAE)评估显着性建议的质量。它采用图像前景的潜在表示,并以鉴别的方式自动编码前景一致性和前景背景的独特性。得到的模型,启用SAE的融合(SAEF),可以组合多个显着性建议,从而产生更可靠的显着性图。在第二阶段,由于融合常常导致过度平滑的显着性图,我们开发自训练的卷积神经网络(STCNN)来缓解这种负面影响。 STCNN将SAEF产生的显着图作为输入。它将信息从高信任的区域传播到低信心的区域。在传播期间,蒸馏出特征表示,导致更尖锐和更好的共同显着性图。我们的方法是全面评估的三个基准,包括MSRC,ICOSEG和COSAL2015,并对最先进的方式表现。此外,我们证明我们的方法可以应用于对象共分割和对象共定位,实现两个应用中的最先进的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号