首页> 外文学位 >Optimization-based summarization and indexing of extended videos, with application to instructional video semantics.
【24h】

Optimization-based summarization and indexing of extended videos, with application to instructional video semantics.

机译:基于优化的扩展视频摘要和索引,应用于教学视频语义。

获取原文
获取原文并翻译 | 示例

摘要

We develop and demonstrate methods to summarize and index extended videos in a semantically informed way, and present semantic compression and linking techniques for instructional videos. We also demonstrate for the first time the extreme degrees to which instructional videos can be usefully compressed and indexed.; We first define and design a generic uniform approach for the selection of key frame subsets from videos. By converting the problem of video summarization into the problem of recognizing those key frame subsequences that optimize pre-defined criteria, we present novel off-line optimization-based approaches, and evaluate their superior results compared to those of existing algorithms. Two separate video summarization semantic criteria are provided and explored, one more favorable for video content summarization, and the other more favorable for network transmission and reconstruction. We further extend off-line optimization approaches to real-time on-line dynamic semantic compression by using a human memory buffer model of time-constrained video perception. This model provides a dynamically ratio-adjustable semantic compression for video summarization and streaming. An entire hierarchy of key frame subsets can be retrieved at no additional cost. We show the relationship between the off-line model and the on-line model, and analyze their performance results.; Next, we show that these methods naturally lead to efficient summarization of instructional videos by defining semantic measures that capture the content of this genre. Focusing on two dominant presentation formats of instructional videos, those of hand written slide and blackboard, we provide a rule-based approach for extracting clean summary frames with significant content. We demonstrate the extraction of “content panorama slides” that summarize hand-written content in these videos, by extracting and stitching together their content while maintaining the qualitative spatial relationship of the text lines and figures. We also introduce a novel concept of “semantic teaching unit”, a meaningful spatial-temporal unit in instructional videos, and present a rule-based method to extract them. We create from semantic teaching units a highly a compact summary (temporally compressed several thousand times), enabling reconstruction and indexing of instructional video content. Finally, we demonstrate how videos containing presentations using formatted computer slides can be efficiently recognized and indexed to their computer-readable sources.
机译:我们开发和演示了以语义知悉的方式总结和索引扩展视频的方法,并介绍了教学视频的语义压缩和链接技术。我们还首次展示了教学视频可以有效压缩和索引的极限程度。我们首先定义和设计一种通用的统一方法,用于从视频中选择关键帧子集。通过将视频汇总问题转换为识别优化预定义标准的关键帧子序列的问题,我们提出了基于离线优化的新颖方法,并评估了它们与现有算法相比的优越结果。提供和探索了两个单独的视频摘要语义标准,一个更有利于视频内容摘要,另一个更有利于网络传输和重构。通过使用受时间限制的视频感知的人类内存缓冲区模型,我们进一步将离线优化方法扩展到实时在线动态语义压缩。该模型为视频摘要和流提供了动态比率可调的语义压缩。关键帧子集的整个层次结构可以免费检索。我们展示了离线模型和在线模型之间的关系,并分析了它们的性能结果。接下来,我们将展示这些方法通过定义捕获该类型内容的语义度量来自然地有效指导视频摘要。针对两种主要的教学视频演示格式,即手写幻灯片和黑板上的演示格式,我们提供了一种基于规则的方法来提取具有重要内容的简洁摘要框架。我们演示了提取“内容全景幻灯片”的过程,这些摘要概述了这些视频中的手写内容,方法是提取并缝合它们的内容,同时保持文本线条和图形的定性空间关系。我们还介绍了“语义教学单元”的新概念,即教学视频中有意义的时空单元,并提出了一种基于规则的方法来提取它们。我们从语义教学单元中创建了一个高度紧凑的摘要(临时压缩了数千次),从而可以对教学视频内容进行重建和索引。最后,我们演示了如何有效识别包含使用格式化的计算机幻灯片的演示文稿的视频并将其索引到其计算机可读源。

著录项

  • 作者

    Liu, Tiecheng.;

  • 作者单位

    Columbia University.;

  • 授予单位 Columbia University.;
  • 学科 Computer Science.
  • 学位 Ph.D.
  • 年度 2003
  • 页码 174 p.
  • 总页数 174
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类 自动化技术、计算机技术;
  • 关键词

  • 入库时间 2022-08-17 11:44:51

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号