首页> 外文会议>IEEE Winter Conference on Applications of Computer Vision >Toward Interactive Self-Annotation For Video Object Bounding Box: Recurrent Self-Learning And Hierarchical Annotation Based Framework
【24h】

Toward Interactive Self-Annotation For Video Object Bounding Box: Recurrent Self-Learning And Hierarchical Annotation Based Framework

机译:面向视频对象边界框的交互式自注释:基于递归自学习和分层注释的框架

获取原文

摘要

Amount and variety of training data drastically affect the performance of CNNs. Thus, annotation methods are becoming more and more critical to collect data efficiently. In this paper, we propose a simple yet efficient Interactive Self-Annotation framework to cut down both time and human labor cost for video object bounding box annotation. Our method is based on recurrent self-supervised learning and consists of two processes: automatic process and interactive process, where the automatic process aims to build a supported detector to speed up the interactive process. In the Automatic Recurrent Annotation, we let an off-the-shelf detector watch unlabeled videos repeatedly to reinforce itself automatically. At each iteration, we utilize the trained model from the previous iteration to generate better pseudo ground-truth bounding boxes than those at the previous iteration, recurrently improving self-supervised training the detector. In the Interactive Recurrent Annotation, we tackle the human-in-the-loop annotation scenario where the detector receives feedback from the human annotator. To this end, we propose a novel Hierarchical Correction module, where the annotated frame-distance binarizedly decreases at each time step, to utilize the strength of CNN for neighbor frames. Experimental results on various video datasets demonstrate the advantages of the proposed framework in generating high-quality annotations while reducing annotation time and human labor costs.
机译:训练数据的数量和种类极大地影响了CNN的性能。因此,注释方法对于有效地收集数据变得越来越重要。在本文中,我们提出了一个简单而有效的交互式自注释框架,以减少视频对象边界框注释的时间和人工成本。我们的方法基于循环式自我监督学习,包括两个过程:自动过程和交互过程,其中自动过程旨在构建一个受支持的检测器以加快交互过程。在自动循环注释中,我们让现成的检测器反复观看未标记的视频以自动增强自身。在每次迭代中,我们利用前一次迭代中的训练模型来生成比前一次迭代中更好的伪地面真假边界框,从而不断改善对检测器的自我监督训练。在“交互式循环注释”中,我们解决了检测者从人工注释者接收反馈的“人在回路”注释场景。为此,我们提出了一种新颖的分层校正模块,其中带注释的帧距离在每个时间步均以二进制方式减小,以利用CNN的强度用于相邻帧。在各种视频数据集上的实验结果证明了提出的框架在生成高质量注释时的优势,同时减少了注释时间和人工成本。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号