首页> 外文期刊>ACM transactions on multimedia computing communications and applications >Semantic Feature Mining for Video Event Understanding
【24h】

Semantic Feature Mining for Video Event Understanding

机译:用于视频事件理解的语义特征挖掘

获取原文
获取原文并翻译 | 示例
           

摘要

Content-based video understanding is extremely difficult due to the semantic gap between low-level vision signals and the various semantic concepts (object, action, and scene) in videos. Though feature extraction from videos has achieved significant progress, most of the previous methods rely only on low-level features, such as the appearance and motion features. Recently, visual-feature extraction has been improved significantly with machine-learning algorithms, especially deep learning. However, there is still not enough work focusing on extracting semantic features from videos directly. The goal of this article is to adopt unlabeled videos with the help of text descriptions to learn an embedding function, which can be used to extract more effective semantic features from videos when only a few labeled samples are available for video recognition. To achieve this goal, we propose a novel embedding convolutional neural network (ECNN). We evaluate our algorithm by comparing its performance on three challenging benchmarks with several popular state-of-the-art methods. Extensive experimental results show that the proposed ECNN consistently and significantly outperforms the existing methods.
机译:由于低级视觉信号与视频中各种语义概念(对象,动作和场景)之间的语义鸿沟,基于内容的视频理解非常困难。尽管从视频中提取特征已经取得了显着进展,但是大多数以前的方法仅依赖于低级特征,例如外观和运动特征。最近,视觉特征提取已通过机器学习算法(尤其是深度学习)得到了显着改善。但是,仍然没有足够的工作来直接从视频中提取语义特征。本文的目的是在文本描述的帮助下采用未标记的视频以学习嵌入功能,当只有少数标记的样本可用于视频识别时,该功能可用于从视频中提取更有效的语义特征。为了实现这个目标,我们提出了一种新颖的嵌入卷积神经网络(ECNN)。我们通过将其在三个具有挑战性的基准上的性能与几种流行的最新方法进行比较来评估我们的算法。大量的实验结果表明,提出的ECNN始终且显着优于现有方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号