首页> 外文期刊>International Journal of Intelligent Systems >Online human action recognition with spatial and temporal skeleton features using a distributed camera network
【24h】

Online human action recognition with spatial and temporal skeleton features using a distributed camera network

机译:使用分布式摄像机网络与空间和时间骨架特征在线人类行动识别

获取原文
获取原文并翻译 | 示例
       

摘要

Online action recognition is an important task for human-centered intelligent services. However, it remains a highly challenging problem due to the high varieties and uncertainties of spatial and temporal scales of human actions. In this paper, the following core ideas are proposed to deal with the online action recognition problem. First, we combine spatial and temporal skeleton features to represent human actions, which include not only geometrical features, but also multiscale motion features, such that both spatial and temporal information of the actions are covered. We use an efficient one-dimensional convolutional neural network to fuse spatial and temporal features and train them for action recognition. Second, we propose a group sampling method to combine the previous action frames and current action frames, which are based on the hypothesis that the neighboring frames are largely redundant, and the sampling mechanism ensures that the long-term contextual information is also considered. Third, the skeletons from multiview cameras are fused in a distributed manner, which can improve the human pose accuracy in the case of occlusions. Finally, we propose a Restful style based client-server service architecture to deploy the proposed online action recognition module on the remote server as a public service, such that camera networks for online action recognition can benefit from this architecture due to the limited onboard computational resources. We evaluated our model on the data sets of JHMDB and UT-Kinect, which achieved highly promising accuracy levels of 80.1% and 96.9%, respectively. Our online experiments show that our memory group sampling mechanism is far superior to the traditional sliding window.
机译:在线行动认可是以人为本的智能服务的重要任务。然而,由于人类行为的空间和时间尺度的高品种和不确定性,它仍然是一个非常具有挑战性的问题。在本文中,提出了以下核心想法来处理在线行动识别问题。首先,我们组合空间和时间骨架特征来代表人类的行为,其不仅包括几何特征,还包括多尺度运动特征,使得覆盖了动作的空间和时间信息。我们使用高效的一维卷积神经网络来熔断空间和时间特征,并培训它们进行动作识别。其次,我们提出了一种组采样方法来组合先前的动作帧和当前动作帧,其基于相邻帧在很大程度上是冗余的假设,并且采样机制确保也考虑了长期上下文信息。第三,来自多视图摄像机的骷髅以分布式方式融合,可以在闭塞的情况下提高人类姿势精度。最后,我们提出了一种基于宁静的风格的客户端 - 服务器服务架构,可以在远程服务器上部署所提出的在线动作识别模块作为公共服务,使得在线动作识别的相机网络可以从此架构中受益于由于车载计算资源有限。我们在JHMDB和UT-Kinect的数据集上评估了我们的模型,其分别实现了高度有希望的准确度和96.9%。我们的在线实验表明,我们的记忆组采样机制远远优于传统的滑动窗口。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号