首页> 外文会议>IEEE International Conference on Multimedia and Expo >Spatial Mask ConvLSTM Network and Intra-Class Joint Training Method for Human Action Recognition in Video
【24h】

Spatial Mask ConvLSTM Network and Intra-Class Joint Training Method for Human Action Recognition in Video

机译:视频中人类动作识别的空间掩模ConvLSTM网络和类内联合训练方法

获取原文

摘要

For action recognition, attention model is widely used, but most of them lack consideration of the relationship of spatial and temporal information. We thus propose a Spatial Mask ConvLSTM Network (SM_ConvLSTM-Net) to determine the attention score of each pixel position. SM_ConvLSTM-Net is used to combine the information of space and time for getting more precise spatial mask, which has a long receptive field in time domain. Furthermore, to combine the connection of different samples from same category, a novel training method called intra-class joint training method is proposed to make network extract the common characteristics related to actions of the same class in different background. Extensive experiments illustrate the effectiveness of our method and our method significantly outperforms the baseline C3D network on UCF101 and HMDB51. Moreover, our approach achieves the best performance on UCF101 and a compared result on HMDB51 in comparison to some state-of-the-art approaches with RGB input.
机译:对于行动识别,注意模型被广泛使用,但大多数人缺乏对空间和时间信息的关系的思考。因此,我们提出了一种空间掩模GUNMLSTM网络(SM_CONVLSTM-NET),以确定每个像素位置的注意得分。 SM_CONVLSTM-NET用于将空间和时间的信息组合,以获取更精确的空间掩码,其在时域中具有长度接收字段。此外,为了组合来自相同类别的不同样本的连接,提出了一种称为类内联训练方法的新颖训练方法,以使网络提取与不同背景中同一类的动作相关的共同特征。广泛的实验说明了我们的方法的有效性,我们的方法显着优于UCF101和HMDB51上的基线C3D网络。此外,我们的方法可以实现UCF101上的最佳性能,与HMDB51的比较结果相比,与RGB输入的某些最先进的方法相比。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号