首页> 外文会议>European conference on computer vision >GeThR-Net: A Generalized Temporally Hybrid Recurrent Neural Network for Multimodal Information Fusion
【24h】

GeThR-Net: A Generalized Temporally Hybrid Recurrent Neural Network for Multimodal Information Fusion

机译:Gethr-net:用于多峰信息融合的通用时间混合复发性神经网络

获取原文

摘要

Data generated from real world events are usually temporal and contain multimodal information such as audio, visual, depth, sensor etc. which are required to be intelligently combined for classification tasks. In this paper, we propose a novel generalized deep neural network architecture where temporal streams from multiple modalities are combined. There are total M+1 (M is the number of modalities) components in the proposed network. The first component is a novel temporally hybrid Recurrent Neural Network (RNN) that exploits the complimentary nature of the multimodal temporal information by allowing the network to learn both modality specific temporal dynamics as well as the dynamics in a multimodal feature space. M additional components are added to the network which extract discriminative but non-temporal cues from each modality. Finally, the predictions from all of these components are linearly combined using a set of automatically learned weights. We perform exhaustive experiments on three different datasets spanning four modalities. The proposed network is relatively 3.5%, 5.7% and 2% better than the best performing temporal multimodal baseline for UCF-101, CCV and Multimodal Gesture datasets respectively.
机译:从真实世界事件生成的数据通常是时间的,并且包含多模式信息,例如音频,视觉,深度,传感器等,其需要智能地组合分类任务。在本文中,我们提出了一种新的广义深度神经网络架构,其中组合了来自多种方式的时间流。在所提出的网络中总共M + 1(M是模态数量)组件。第一组件是一种新的时间混合复发性神经网络(RNN),其通过允许网络来学习模态特定的时间动态以及多模式特征空间中的动态来利用多式联运时间信息的互补性。 M附加组件被添加到网络中提取来自每个模态的判别但非时间线索。最后,使用一组自动学习权重来线性地组合这些组件的预测。我们在跨越四种方式的三个不同数据集中执行详尽的实验。所提出的网络相对3.5%,5.7%和2%比分别UCF-101,CCV和多模态数据集手势最好进行时间多峰基线更好。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号