首页> 外文会议>International Conference on Emerging Technologies >A Deep Learning based Approach for Precise Video Tagging
【24h】

A Deep Learning based Approach for Precise Video Tagging

机译:基于深度学习的精确视频标记方法

获取原文

摘要

With the increase in smart devices and abundance of video contents, efficient techniques for the indexing, analysis and retrieval of videos are becoming more and more desirable. Improved indexing and automated analysis of millions of videos could be accomplished by getting videos tagged automatically. A lot of existing methods fail to precisely tag videos because of their lack of ability to capture the video context. The context in a video represents the interactions of objects in a scene and their overall meaning. In this work, we propose a novel approach that integrates the video scene ontology with CNN (Convolutional Neural Network) for improved video tagging. Our method captures the content of a video by extracting the information from individual key frames. The key frames are then fed to a CNN based deep learning model to train its parameters. The trained parameters are used to generate the most frequent tags. Highly frequent tags are used to summarize the input video. The proposed technique is benchmarked on the most widely used dataset of video activities, namely, UCF-101. Our method managed to achieve an overall accuracy of 99.8% with an F1- score of 96.2%.
机译:随着智能设备的增加和视频内容的丰富,用于视频的索引,分析和检索的有效技术变得越来越受欢迎。通过自动标记视频,可以改进对数百万个视频的索引编制和自动分析。由于缺乏捕获视频上下文的能力,许多现有方法无法精确标记视频。视频中的上下文表示场景中对象的交互及其整体含义。在这项工作中,我们提出了一种新颖的方法,该方法将视频场景本体与CNN(卷积神经网络)集成在一起,以改进视频标记。我们的方法通过从各个关键帧中提取信息来捕获视频的内容。然后将关键帧馈送到基于CNN的深度学习模型以训练其参数。训练有素的参数用于生成最频繁的标签。频繁使用的标签用于汇总输入视频。所提议的技术以最广泛使用的视频活动数据集UCF-101为基准。我们的方法设法达到99.8%的整体准确率和96.2%的F1分数。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号