首页> 外文会议>International symposium on visual computing >Leveraging Multi-modal Analyses and Online Knowledge Base for Video Aboutness Generation
【24h】

Leveraging Multi-modal Analyses and Online Knowledge Base for Video Aboutness Generation

机译:利用多模式分析和在线知识库生成视频信息

获取原文

摘要

The Internet has a huge volume of unlabeled videos from diverse sources, making it difficult for video providers to organize and for viewers to consume the content. This paper defines the problem of video aboutness generation (i.e., the automatic generation of a concise natural-language description about a video) and characterizes its differences from closely related problems such as video summarization and video caption. We then made an attempt to provide a solution to this problem. Our proposed system exploits multi-modal analyses of audio, text and visual content of the video and leverages the Internet to identify a top-matched aboutness description. Through an exploratory study involving human judges evaluating a variety of test videos, we found support of the proposed approach.
机译:互联网上有大量来自不同来源的未标记视频,这使得视频提供商难以组织,观看者也难以使用这些内容。本文定义了视频相关性生成的问题(即,自动生成有关视频的简洁自然语言描述),并描述了其与紧密相关的问题(如视频摘要和视频字幕)的区别。然后,我们尝试为该问题提供解决方案。我们提出的系统利用视频的音频,文本和视觉内容的多模式分析,并利用Internet来识别最匹配的关于状态的描述。通过一项涉及人类法官评估各种测试视频的探索性研究,我们发现了所提议方法的支持。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号