首页> 外文会议>IEEE International Conference on Image Processing >Spatial-Temporal Data Augmentation Based on LSTM Autoencoder Network for Skeleton-Based Human Action Recognition
【24h】

Spatial-Temporal Data Augmentation Based on LSTM Autoencoder Network for Skeleton-Based Human Action Recognition

机译:基于LSTM自动编码器网络的时空数据增强用于基于骨骼的人体动作识别

获取原文

摘要

Data augmentation is known to be of crucial importance for the generalization of RNN-based methods of skeleton-based human action recognition. Traditional data augmentation methods artificially adopt various transformations merely in spatial domain, which lack effective temporal representation. This paper extends traditional Long Short-Term Memory (LSTM) and presents a novel LSTM autoencoder network (LSTM-AE) for spatial-temporal data augmentation. In the LSTM-AE, the LSTM network preserves the temporal information of skeleton sequences, and the autoencoder architecture can automatically eliminate irrelevant and redundant information. Meanwhile, a regularized cross-entropy loss is defined to guide the LSTM-AE to learn more suitable representations of skeleton data. Experimental results on the currently largest NTU RGB+D dataset and public SmartHome dataset verify that the proposed model outperforms the state-of-the-art methods, and can be integrated with most of the RNN-based action recognition models easily.
机译:众所周知,数据增强对于基于RNN的基于骨骼的人类动作识别方法的推广至关重要。传统的数据扩充方法仅在空间域中人为地采用各种变换,而这些变换缺乏有效的时间表示。本文扩展了传统的长期短期记忆(LSTM),并提出了一种用于时空数据扩充的新型LSTM自动编码器网络(LSTM-AE)。在LSTM-AE中,LSTM网络保留了骨架序列的时间信息,并且自动编码器体系结构可以自动消除无关紧要的信息。同时,定义了正则化的交叉熵损失,以指导LSTM-AE学习更合适的骨架数据表示。在当前最大的NTU RGB + D数据集和公共SmartHome数据集上的实验结果证明,所提出的模型优于最新方法,并且可以轻松地与大多数基于RNN的动作识别模型集成。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号