首页> 外文会议>ACM multimedia conference >Learning Frame Relevance for Video Classification
【24h】

Learning Frame Relevance for Video Classification

机译:视频分类的学习框架相关性

获取原文

摘要

Traditional video classification methods typically require a large number of labeled training video frames to achieve satisfactory performance. However, in the real world, we usually only have sufficient labeled video clips (such as tagged online videos) but lack labeled video frames. In this paper, we formalize the video classification problem as a Multi-Instance Learning (MIL) problem, an emerging topic in machine learning in recent years, which only needs bag (video clip) labels. To solve the problem, we propose a novel Parameterized Class-to-Bag (P-C2B) Distance method to learn the relative importance of a training instance with respect to its labeled classes, such that the instance level labeling ambiguity in MIL is tackled and the frame relevances of training video data with respect to the semantic concepts of interest are given. Promising experimental results have demonstrated the effectiveness of the proposed method.
机译:传统的视频分类方法通常需要大量带标签的训练视频帧才能获得令人满意的性能。但是,在现实世界中,我们通常只有足够的带标签的视频剪辑(例如带标签的在线视频),而缺少带标签的视频帧。在本文中,我们将视频分类问题形式化为多实例学习(MIL)问题,这是近年来机器学习中一个新兴的话题,它仅需要标签(视频剪辑)即可。为了解决该问题,我们提出了一种新颖的参数化类到袋(P-C2B)距离方法,以学习训练实例相对于其标记类的相对重要性,从而解决了MIL中实例级别的标记歧义并给出了关于感兴趣的语义概念的训练视频数据的帧相关性。有希望的实验结果证明了该方法的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号