首页> 外文期刊>IEEE transactions on visualization and computer graphics >Stage-wise Salient Object Detection in 360° Omnidirectional Image via Object-level Semantical Saliency Ranking
【24h】

Stage-wise Salient Object Detection in 360° Omnidirectional Image via Object-level Semantical Saliency Ranking

机译:通过对象级语义显着性排名在360°全向图像中检测阶段明显的物体检测

获取原文
获取原文并翻译 | 示例
       

摘要

The 2D image based salient object detection (SOD) has been extensively explored, while the 360° omnidirectional image based SOD has received less research attention and there exist three major bottlenecks that are limiting its performance. Firstly, the currently available training data is insufficient for the training of 360° SOD deep model. Secondly, the visual distortions in 360° omnidirectional images usually result in large feature gap between 360° images and 2D images; consequently, the widely used stage-wise training—a widely-used solution to alleviate the training data shortage problem, becomes infeasible when conducing SOD in 360° omnidirectional images. Thirdly, the existing 360° SOD approach has followed a multi-task methodology that performs salient object localization and segmentation-like saliency refinement at the same time, being faced with extremely large problem domain, making the training data shortage dilemma even worse. To tackle all these issues, this paper divides the 360° SOD into a multi-staqe task, the key rationale of which is to decompose the original complex problem domain into sequential easy sub problems that only demand for small-scale training data. Meanwhile, we learn how to rank the “object-level semantical saliency”, aiming to locate salient viewpoints and objects accurately. Specifically, to alleviate the training data shortage problem, we have released a novel dataset named 360-SSOD, containing 1,105 360° omnidirectional images with manually annotated object-level saliency ground truth, whose semantical distribution is more balanced than that of the existing dataset. Also, we have compared the proposed method with 13 SOTA methods, and all quantitative results have demonstrated the performance superiority.
机译:基于2D图像的显着对象检测(SOD)已被广泛探索,而基于360°的全向图像的SOD已经收到了较少的研究人身,并且存在限制其性能的三个主要瓶颈。首先,目前可用的培训数据不足以培训360°SOD深度模型。其次,在360°的全向图像中的视觉扭曲通常导致360°图像和2D图像之间的大特征间隙;因此,广泛使用的阶段明智的训练 - 一种广泛使用的解决方案,以缓解训练数据短缺问题,在360°的全向图像中导致SOD时变得不可行。第三,现有的360°SOD方法遵循了一个多任务方法,同时表现出突出的对象定位和分段式显着细化,面临极大的问题域,使训练数据短缺困境甚至更糟糕。为了解决所有这些问题,本文将360°SOD划分为多节奏任务,其关键基本基本原理是将原始复杂问题域分解为只需要对小规模培训数据的顺序易次级问题。同时,我们学习如何对“对象级语义显着性”进行排序,旨在准确定位突出视点和对象。具体而言,为了减轻培训数据短缺问题,我们发布了一个名为360-SSOD的新型数据集,其中包含1,105个360°的全向图像,具有手动注释的对象级显着性地面真理,其语义分布比现有数据集更加平衡。此外,我们已经将提出的方法与13种SOTA方法进行了比较,并且所有定量结果都证明了性能优势。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号