首页> 外文期刊>Neural computing & applications >Spatiotemporal saliency-based multi-stream networks with attention-aware LSTM for action recognition
【24h】

Spatiotemporal saliency-based multi-stream networks with attention-aware LSTM for action recognition

机译:基于时空显着性的多流网络,注意感知LSTM用于动作识别

获取原文
获取原文并翻译 | 示例
获取外文期刊封面目录资料

摘要

Human action recognition is a process of labeling video frames with action labels. It is a challenging research topic since the background of videos is usually chaotic, which will reduce the performance of traditional human action recognition methods. In this paper, we propose a novel spatiotemporal saliency-based multi-stream ResNets (STS), which combines three streams (i.e., a spatial stream, a temporal stream and a spatiotemporal saliency stream) for human action recognition. Further, we propose a novel spatiotemporal saliency-based multi-stream ResNets with attention-aware long short-term memory (STS-ALSTM) network. The proposed STS-ALSTM model combines deep convolutional neural network (CNN) feature extractors with three attention-aware LSTMs to capture the temporal long-term dependency relationships between consecutive video frames, optical flow frames or spatiotemporal saliency frames. Experimental results on UCF-101 and HMDB-51 datasets demonstrate that our proposed STS method and STS-ALSTM model obtain competitive performance compared with the state-of-the-art methods.
机译:人类行动识别是用动作标签标记视频框架的过程。这是一个具有挑战性的研究主题,因为视频背景通常是混乱的,这将降低传统人类行动识别方法的性能。在本文中,我们提出了一种新的基于时空显着性的多流Emenet(STS),其结合了用于人类动作识别的三个流(即空间流,时间流和时空性效力流和时空性瞬时性流。此外,我们提出了一种具有注意力感知的长短期存储器(STS-ALSTM)网络的新型时空显着的多流ESNET。所提出的STS-Alstm模型结合了深度卷积神经网络(CNN)特征提取器,具有三个注意力感知的LSTM,以捕获连续视频帧,光学流帧或时空显着帧之间的时间长期依赖关系。 UCF-101和HMDB-51数据集上的实验结果表明,与最先进的方法相比,我们所提出的STS方法和STS-Alstm模型获得了竞争性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号