首页> 外文会议>IEEE Conference on Computer Vision and Pattern Recognition Workshops >Multimodal Multi-Stream Deep Learning for Egocentric Activity Recognition
【24h】

Multimodal Multi-Stream Deep Learning for Egocentric Activity Recognition

机译:多峰多流深入学习,用于自我监测活动识别

获取原文

摘要

In this paper, we propose a multimodal multi-stream deep learning framework to tackle the egocentric activity recognition problem, using both the video and sensor data. First, we experiment and extend a multi-stream Convolutional Neural Network to learn the spatial and temporal features from egocentric videos. Second, we propose a multistream Long Short-Term Memory architecture to learn the features from multiple sensor streams (accelerometer, gyroscope, etc.). Third, we propose to use a two-level fusion technique and experiment different pooling techniques to compute the prediction results. Experimental results using a multimodal egocentric dataset show that our proposed method can achieve very encouraging performance, despite the constraint that the scale of the existing egocentric datasets is still quite limited.
机译:在本文中,我们提出了一种多模式多流深入学习框架来解决视频和传感器数据的精神度活动识别问题。首先,我们尝试并扩展多流卷积神经网络,以了解从Egentric视频的空间和时间特征。其次,我们提出了一种多阵线长短期内存架构,以了解来自多个传感器流(加速度计,陀螺仪等)的特征。第三,我们建议使用双层融合技术并实验不同的汇集技术来计算预测结果。实验结果采用多模式的自主特征数据集显示我们所提出的方法可以实现非常令人鼓舞的性能,尽管存在现有的EnoCentric数据集的规模仍然相当有限。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号