首页> 外文期刊>IEEE Transactions on Circuits and Systems for Video Technology >Spatial and Motion Saliency Prediction Method Using Eye Tracker Data for Video Summarization
【24h】

Spatial and Motion Saliency Prediction Method Using Eye Tracker Data for Video Summarization

机译:使用眼跟踪器数据进行视频汇总的空间和运动耐药性预测方法

获取原文
获取原文并翻译 | 示例
获取外文期刊封面目录资料

摘要

Video summarization is the process to extract the most significant contents of a video and to represent it in a concise form. The existing methods for video summarization could not achieve a satisfactory result for a video with camera movement and significant illumination changes. To solve these problems, in this paper, a new framework for video summarization is proposed based on eye tracker data, as human eyes can track moving object accurately in these cases. The smooth pursuit is the state of eye movement when a user follows a moving object in a video. This motivates us to implement a new method to distinguish smooth pursuit from other type of gaze points, such as fixation and saccade. The smooth pursuit provides only the location of moving objects in a video frame; however, it does not indicate whether the located moving objects are very attractive (i.e., salient regions) to viewers or not, as well as the amount of motion of the moving objects. The amount of salient regions and object motions are the two important features to measure the viewer's attention level for determining the key frames for video summarization. To find the most attractive objects, a new spatial saliency prediction method is also proposed by constructing a saliency map around each smooth pursuit gaze point based on human visual field, such as fovea, parafoveal, and perifovea regions. To identify the amount of object motions, the total distances between the current and the previous gaze points of viewers during smooth pursuit are measured as a motion saliency score. The motivation is that the movement of eye gaze is related to the motion of the objects during smooth pursuit. Finally, both spatial and motion saliency maps are combined to obtain an aggregated saliency score for each frame and a set of key frames are selected based on user selected or system default skimming ratio. The proposed method is implemented on Office video data set that contains videos with camera movements and illumination changes. Experimental results confirm the superior performance of the proposed spatial and motion saliency prediction method compared with the state-of-the-art methods.
机译:视频摘要是提取视频最重要内容的过程并以简洁的形式表示。用于视频摘要的现有方法无法实现相机运动和显着的照明变化的视频的令人满意的结果。为了解决这些问题,本文基于眼睛跟踪器数据提出了一种用于视频摘要的新框架,因为人眼可以在这些情况下准确地跟踪移动物体。当用户在视频中遵循移动对象时,平滑追踪是眼球运动状态。这使我们能够实施一种新的方法,以区分从其他类型的凝视点,例如固定和扫视。平滑追求仅提供视频帧中的移动物体的位置;然而,它并不意味着所定位的移动物体是非常有吸引力的(即,显着区域),以及移动物体的运动量。突出区域和对象运动的量是测量观众的注意力水平,以确定用于确定视频概括的关键帧的两个重要功能。为了找到最具吸引力的物体,还通过构建基于人类视野的每个平滑追踪凝视点周围的显着性图,例如FOVEA,PARAFOVEAL和PERIFOVEA地区来提出新的空间显着性预测方法。为了识别物体运动量,测量在平滑追踪期间观察者的电流与先前凝视点之间的总距离被测量为运动显着分数。动机是眼睛凝视的运动与在平滑追踪过程中对象的运动有关。最后,组合空间和运动显着性图以获得每个帧的聚合显着性分数,并且基于用户选择或系统默认扫描比选择一组关键帧。所提出的方法在办公室视频数据集上实现,其中包含具有相机移动和照明变化的视频。实验结果证实了与最先进的方法相比,所提出的空间和运动显着性预测方法的优越性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号