首页> 外文期刊>Neurocomputing >Deep clustering for weakly-supervised semantic segmentation in autonomous driving scenes
【24h】

Deep clustering for weakly-supervised semantic segmentation in autonomous driving scenes

机译:深度聚类用于自动驾驶场景中的弱监督语义分割

获取原文
获取原文并翻译 | 示例
       

摘要

Weakly-supervised semantic segmentation (WSSS) using only tags can significantly ease the label costing, because full supervision needs pixel-level labeling. It is, however, a very challenging task because it is not straightforward to associate tags to visual appearance. Existing researches can only do tag-based WSSS on simple images, where only two or three tags exist in each image, and different images usually have different tags, such as the PASCAL VOC dataset. Therefore, it is easy to relate the tags to visual appearance and supervise the segmentation. However, real-world scenes are much more complex. Especially, the autonomous driving scenes usually contain nearly 20 tags in each image and those tags can repetitively appear from image to image, which means the existing simple image strategy does not work. In this paper, we propose to solve the problem by using region based deep clustering. The key idea is that, since each tagged object is repetitively appearing from image to image, it allows us to find the common appearance through region clustering, and particular deep neural network based clustering. Later, we relate the clustered region appearance to tags and utilize the tags to supervise the segmentation. Furthermore, regions found by clustering with weak supervision can be very noisy. We further propose a mechanic to improve and refine the supervision in an iterative manner. To our best knowledge, it is the first time that image tags weakly-supervised semantic segmentation can be applied in complex autonomous driving datasets with still images. Experimental results on the Cityscapes and CamVid datasets demonstrate the effectiveness of our method. (C) 2019 Elsevier B.V. All rights reserved.
机译:仅使用标签的弱监督语义分段(WSSS)可以大大减轻标签成本,因为全面监督需要像素级标签。然而,这是一项非常具有挑战性的任务,因为将标签与视觉外观相关联并不容易。现有研究只能在简单图像上执行基于标签的WSSS,其中每个图像中仅存在两个或三个标签,并且不同的图像通常具有不同的标签,例如PASCAL VOC数据集。因此,很容易将标签与视觉外观相关联并监督分割。但是,现实世界的场景要复杂得多。尤其是,自动驾驶场景通常在每个图像中包含近20个标签,并且这些标签可以在图像之间重复出现,这意味着现有的简单图像策略不起作用。在本文中,我们建议通过使用基于区域的深度聚类来解决该问题。关键思想是,由于每个标记的对象在图像之间反复出现,因此它使我们能够通过区域聚类(尤其是基于深度神经网络的聚类)找到共同的外观。稍后,我们将聚类的区域外观与标签相关联,并利用标签来监督细分。此外,在监督不力的情况下通过聚类发现的区域可能非常嘈杂。我们进一步提出一种机制,以迭代的方式改进和完善监管。据我们所知,这是第一次将图像标签弱监督语义分割应用于具有静态图像的复杂自动驾驶数据集中。在Cityscapes和CamVid数据集上的实验结果证明了我们方法的有效性。 (C)2019 Elsevier B.V.保留所有权利。

著录项

  • 来源
    《Neurocomputing》 |2020年第14期|20-28|共9页
  • 作者

  • 作者单位

    Tencent Res Beijing Peoples R China|Tsinghua Univ Beijing Peoples R China;

    Univ Sci & Technol Beijing Inst Artificial Intelligence Beijing Peoples R China;

    Univ Amsterdam Amsterdam Netherlands;

  • 收录信息 美国《科学引文索引》(SCI);美国《工程索引》(EI);
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Weak supervision; Semantic segmentation; Deep clustering; Autonomous driving;

    机译:监管薄弱;语义分割;深度集群;自动驾驶;

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号