首页> 外文会议>2018 IEEE Conference on Multimedia Information Processing and Retrieval >Sequential Deep Learning for Disaster-Related Video Classification
【24h】

Sequential Deep Learning for Disaster-Related Video Classification

机译:顺序深度学习用于与灾难相关的视频分类

获取原文
获取原文并翻译 | 示例

摘要

Videos serve to convey complex semantic information and ease the understanding of new knowledge. However, when mixed semantic meanings from different modalities (i.e., image, video, text) are involved, it is more difficult for a computer model to detect and classify the concepts (such as flood, storm, and animals). This paper presents a multimodal deep learning framework to improve video concept classification by leveraging recent advances in transfer learning and sequential deep learning models. Long Short-Term Memory (LSTM) Recurrent Neural Networks (RNN) models are then used to obtain the sequential semantics for both audio and textual models. The proposed framework is applied to a disaster-related video dataset that includes not only disaster scenes, but also the activities that took place during the disaster event. The experimental results show the effectiveness of the proposed framework.
机译:视频可传达复杂的语义信息并简化对新知识的理解。但是,当涉及来自不同形式(即图像,视频,文本)的混合语义时,计算机模型更难检测和分类概念(例如洪水,暴风雨和动物)。本文提出了一种多模式深度学习框架,以利用转移学习和顺序深度学习模型的最新进展来改进视频概念分类。然后,使用长短期记忆(LSTM)递归神经网络(RNN)模型来获得音频和文本模型的顺序语义。所提出的框架被应用于与灾难有关的视频数据集,该数据集不仅包括灾难现场,还包括灾难事件期间发生的活动。实验结果表明了该框架的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号