首页> 外文会议>Asian Conference on Computer Vision >Unseen Object Segmentation in Videos via Transferable Representations
【24h】

Unseen Object Segmentation in Videos via Transferable Representations

机译:通过可转移表示形式的视频中看不见的对象分割

获取原文

摘要

In order to learn object segmentation models in videos, conventional methods require a large amount of pixel-wise ground truth annotations. However, collecting such supervised data is time-consuming and labor-intensive. In this paper, we exploit existing annotations in source images and transfer such visual information to segment videos with unseen object categories. Without using any annotations in the target video, we propose a method to jointly mine useful segments and learn feature representations that better adapt to the target frames. The entire process is decomposed into two tasks: (1) solving a submodular function for selecting object-like segments, and (2) learning a CNN model with a transferable module for adapting seen categories in the source domain to the unseen target video. We present an iterative update scheme between two tasks to self-learn the final solution for object segmentation. Experimental results on numerous benchmark datasets show that the proposed method performs favorably against the state-of-the-art algorithms.
机译:为了学习视频中的对象分割模型,常规方法需要大量的像素级地面真相注释。但是,收集这样的监督数据既费时又费力。在本文中,我们利用了源图像中的现有注释,并将这些视觉信息转移到了具有看不见的对象类别的视频片段上。在目标视频中不使用任何注释的情况下,我们提出了一种方法来联合挖掘有用的片段并学习更好地适应目标帧的特征表示。整个过程分解为两个任务:(1)解决用于选择对象段的亚模块功能;(2)使用可转移模块学习CNN模型,以使源域中的可见类别适应于看不见的目标视频。我们提出了两个任务之间的迭代更新方案,以自学习对象分割的最终解决方案。在大量基准数据集上的实验结果表明,所提出的方法相对于最新算法具有良好的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号