...
首页> 外文期刊>IEEE Transactions on Image Processing >Video Saliency Prediction Using Spatiotemporal Residual Attentive Networks
【24h】

Video Saliency Prediction Using Spatiotemporal Residual Attentive Networks

机译:使用时空残余关节网络的视频显着性预测

获取原文
获取原文并翻译 | 示例
           

摘要

This paper proposes a novel residual attentive learning network architecture for predicting dynamic eye-fixation maps. The proposed model emphasizes two essential issues, i.e., effective spatiotemporal feature integration and multi-scale saliency learning. For the first problem, appearance and motion streams are tightly coupled via dense residual cross connections, which integrate appearance information with multi-layer, comprehensive motion features in a residual and dense way. Beyond traditional two-stream models learning appearance and motion features separately, such design allows early, multi-path information exchange between different domains, leading to a unified and powerful spatiotemporal learning architecture. For the second one, we propose a composite attention mechanism that learns multi-scale local attentions and global attention priors end-to-end. It is used for enhancing the fused spatiotemporal features via emphasizing important features in multi-scales. A lightweight convolutional Gated Recurrent Unit (convGRU), which is flexible for small training data situation, is used for long-term temporal characteristics modeling. Extensive experiments over four benchmark datasets clearly demonstrate the advantage of the proposed video saliency model over other competitors and the effectiveness of each component of our network. Our code and all the results will be available at https://github.com/ashleylqx/STRA-Net.
机译:本文提出了一种用于预测动态眼固定图的新型残差学习网络架构。所提出的模型强调了两个基本问题,<斜体> i.e 。,有效的时空特征集成和多尺度的粘性学习。对于第一个问题,外观和运动流通过密集的残余交叉连接紧密耦合,其以残余和致密的方式将外观信息与多层综合运动特征集成在一起。除了传统的两流模型之外,可以单独学习外观和运动功能,这种设计允许在不同域之间的早期,多路径信息交换,导致统一和强大的时空学习架构。对于第二个,我们提出了一种综合关注机制,这些机制学习多规模的本地关注和全球注意力前端的结束。它用于通过强调多尺度的重要特征来增强熔融的时空特征。用于小型训练数据情况灵活的轻量级卷积栅格复发单元(Concrecru)用于长期时间特性建模。超过四个基准数据集的广泛实验清楚地展示了所提出的视频显着模型对其他竞争对手的优势以及我们网络每个组件的有效性。我们的代码和所有结果将在 https://github.com/ashleylqx/stra-net 上获得。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号