【24h】

Action Anticipation from Multimodal Data

机译:来自多式联算数据的行动预期

获取原文

摘要

The idea of multi-sensor data fusion is to combine the data coming from different sensors to provide more accurate and complementary information to solve a specific task. Our goal is to build a shared representation related to data coming from different domains, such as images, audio signal, heart rate, acceleration, etc., in order to anticipate daily activities of a user wearing multimodal sensors. To this aim, we consider the Stanford-ECM Dataset which contains syncronized data acquired with different sensors: video, acceleration and heart rate signals. The dataset is adapted to our action prediction task by identifying the transitions from the generic "Unknown" class to a specific "Activity". We discuss and compare a Siamese Network with the Multi Layer Perceptron and the 1D CNN where the input is an unknown observation and the output is the next activity to be observed. The feature representations obtained with the considered deep architecture are classified with SVM or KNN classifiers. Experimental results pointed out that prediction from multimodal data seems a feasible task, suggesting that multimodality improves both classification and prediction. Nevertheless, the task of reliably predicting next actions is still open and requires more investigations as well as the availability of multimodal dataset, specifically built for prediction purposes.
机译:多传感器数据融合的思想是将来自不同传感器的数据组合,以提供更准确和互补的信息来解决特定任务。我们的目标是建立与来自不同域的数据相关的共享表示,例如图像,音频信号,心率,加速等,以期望用户佩戴多式联传感器的日常活动。为此目的,我们考虑斯坦福-ECM数据集,其中包含使用不同传感器获取的同步数据:视频,加速度和心率信号。数据集通过将通用“未知”类的转换标识为特定的“活动”来适应我们的动作预测任务。我们将暹罗网络与多层Perceptron和1D CNN讨论并比较,其中输入是未知观察,并且输出是要观察的下一个活动。使用所考虑的深度架构获得的特征表示与SVM或KNN分类器分类。实验结果指出,来自多模式数据的预测似乎是可行的任务,表明多模化改善了分类和预测。尽管如此,可靠地预测下一个措施的任务仍然是开放的,并且需要更多的调查以及多模式数据集的可用性,专门用于预测目的。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号