首页> 外文期刊>IEEE Transactions on Circuits and Systems for Video Technology >Motion-Aware Feature Enhancement Network for Video Prediction
【24h】

Motion-Aware Feature Enhancement Network for Video Prediction

机译:运动感知功能增强网络用于视频预测

获取原文
获取原文并翻译 | 示例
           

摘要

Video prediction is challenging, due to the pixel-level precision requirement and the difficulty in capturing scene dynamics. Most approaches tackle the problems by pixel-level reconstruction objectives and two decomposed branches, which still suffer from blurry generations or dramatic degradations in long-term prediction. In this paper, we propose a Motion-Aware Feature Enhancement (MAFE) network for video prediction to produce realistic future frames and achieve relatively long-term predictions. First, a Channel-wise and Spatial Attention (CSA) module is designed to extract motion-aware features, which enhances the contribution of important motion details during encoding, and subsequently improves the discriminability of attention map for the frame refinement. Second, a Motion Perceptual Loss (MPL) is proposed to guide the learning of temporal cues, which benefits to robust long-term video prediction. Extensive experiments on three human activity video datasets: KTH, Human3.6M, and PennAction demonstrate the effectiveness of the proposed video prediction model compared with the state-of-the-art approaches.
机译:由于像素级精度要求和捕获场景动态的难度,视频预测是具有挑战性的。大多数方法通过像素级重建目标和两个分解的分支来解决问题,这仍然遭受模糊的几代或长期预测中的剧烈降低。在本文中,我们提出了一种动作感知功能增强(MAFE)网络,用于视频预测,以产生逼真的未来帧并实现相对长期的预测。首先,旨在提取通道和空间注意(CSA)模块以提取运动感知特征,这增强了在编码期间的重要运动细节的贡献,随后提高了帧细化的关注图的可怜。其次,提出了运动感知损失(MPL)以指导时间线索的学习,这有利于强大的长期视频预测。对三个人类活动的广泛实验视频数据集:Kth,Human3.6M和Pennaction展示了所提出的视频预测模型的有效性与最先进的方法相比。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号