首页> 外文会议>International Conference on Image and Graphics >Spatial-Temporal Bottom-Up Top-Down Attention Model for Action Recognition
【24h】

Spatial-Temporal Bottom-Up Top-Down Attention Model for Action Recognition

机译:用于行动识别的空间 - 时间自下而上的自视图

获取原文

摘要

Driven by the importance of capturing non-local information in video understanding, we propose Spatial-temporal Bottom-up Top-down Attention Module (STBTA). Features are processed across in multiple scales and then combined to best capture the spatial relationships associated with the region of interest and the surrounding environment in a complicated scene. Attention maps are used for adaptive feature refinement. STBTA can be plugged into any feedforward network architectures and is end-to-end trainable along with CNN. Extensive experiments on UCF101, HMDB51, Kinetics-400 datasets demonstrate that the proposed method can improve the performance for action recognition.
机译:在视频理解中捕获非本地信息的重要性驱动,我们提出了空间 - 时间自下而下的视图模块(StBta)。在多个尺度上处理特征,然后组合以最佳捕获与感兴趣区域和复杂场景中的周围环境相关联的空间关系。注意图用于自适应特征精制。 STBTA可以插入任何前馈网络架构,并与CNN一起结束培训。在UCF101,HMDB51,动力学-400数据集上进行了广泛的实验表明,该方法可以提高动作识别的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号