首页> 外文期刊>IEEE Transactions on Image Processing >Visual Attention Modeling for Stereoscopic Video: A Benchmark and Computational Model
【24h】

Visual Attention Modeling for Stereoscopic Video: A Benchmark and Computational Model

机译:立体视频的视觉注意建模:基准和计算模型

获取原文
获取原文并翻译 | 示例
           

摘要

In this paper, we investigate the visual attention modeling for stereoscopic video from the following two aspects. First, we build one large-scale eye tracking database as the benchmark of visual attention modeling for stereoscopic video. The database includes 47 video sequences and their corresponding eye fixation data. Second, we propose a novel computational model of visual attention for stereoscopic video based on Gestalt theory. In the proposed model, we extract the low-level features, including luminance, color, texture, and depth, from discrete cosine transform coefficients, which are used to calculate feature contrast for the spatial saliency computation. The temporal saliency is calculated by the motion contrast from the planar and depth motion features in the stereoscopic video sequences. The final saliency is estimated by fusing the spatial and temporal saliency with uncertainty weighting, which is estimated by the laws of proximity, continuity, and common fate in Gestalt theory. Experimental results show that the proposed method outperforms the state-of-the-art stereoscopic video saliency detection models on our built large-scale eye tracking database and one other database (DML-ITRACK-3D).
机译:在本文中,我们将从以下两个方面研究立体视频的视觉注意建模。首先,我们建立了一个大规模的眼动追踪数据库,作为立体视频视觉注意力建模的基准。该数据库包括47个视频序列及其对应的眼睛注视数据。其次,我们基于格式塔理论提出了一种新颖的立体视频视觉注意力计算模型。在提出的模型中,我们从离散余弦变换系数中提取出包括亮度,颜色,纹理和深度在内的低级特征,这些低特征用于计算空间显着性的特征对比度。通过根据立体视频序列中的平面和深度运动特征的运动对比度计算时间显着性。最终显着性是通过将空间和时间显着性与不确定性加权相融合来估计的,不确定性权重由格式塔理论中的邻近性,连续性和共同命运定律来估计。实验结果表明,在我们建立的大型眼睛跟踪数据库和另一个数据库(DML-ITRACK-3D)上,该方法优于最新的立体视频显着性检测模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号