首页> 外文学位 >Who moved my slide? Recognizing entities in a lecture video and its applications.
【24h】

Who moved my slide? Recognizing entities in a lecture video and its applications.

机译:谁移动了我的幻灯片?在演讲视频及其应用中识别实体。

获取原文
获取原文并翻译 | 示例

摘要

Lecture videos have proliferated in recent years thanks to the increasing bandwidths of Internet connections and availability of video cameras. Despite the massive volume of videos available, there are very few systems that parse useful information from them. Extracting meaningful data can help with searching and indexing of lecture videos as well as improve understanding and usability for the viewers. While video tags and user preferences are good indicators for relevant videos, it is completely dependent on human-generated data. Furthermore, many lecture videos are technical by nature and sparse video tags are too coarse-grained to relate parts of a video by a specific topic.;While extracting the text from the presentation slide will ameliorate this issue, a lecture video still contains significantly more information than what is just available on the presentation slides. That is, the actions and words of the speaker contribute to a richer and more nuanced understanding of the lecture material. The goal of the Semantically Linked Instructional Content (SLIC) project is to relate videos using more specific and relevant features such as slide text and other entities.;In this work, we will present the algorithms used to recognize the entities of the lecture. Specifically, the entities in lecture videos are the laser and pointing hand gestures and the location of the slide and its text and images in the video. Our algorithms work under the assumption that the slide location (homography) is known for each frame and extend the knowledge of the scene. Specifically, gestures inform when and where on a slide notable events occur.;We will also show how recognition of these entities can help with understanding lectures better and energy-savings on mobile devices. We conducted a user study that shows that magnifying text based on laser gestures on a slide helps direct a viewer's attention to the relevant text. We also performed empirical measurements on real cellphones to conrm that selectively dimming less relevant regions of the video frame would reduce energy consumption significantly.
机译:近年来,由于Internet连接带宽的增加和摄像机的可用性,讲座视频激增了。尽管有大量可用的视频,但很少有系统可以解析其中的有用信息。提取有意义的数据可以帮助对演讲视频进行搜索和索引,以及提高观众的理解力和可用性。视频标签和用户偏好是相关视频的良好指标,但它完全取决于人工生成的数据。此外,许多讲座视频本质上是技术性的,而稀疏的视频标签太粗糙而无法将视频的各个部分与特定主题相关联。;虽然从演示幻灯片中提取文本可以缓解此问题,但讲座视频仍然包含更多内容比演示幻灯片中提供的信息更多。即,说话者的动作和言语有助于对讲座材料的更丰富和更细致的理解。语义链接的教学内容(SLIC)项目的目标是使用更具体且相关的功能(例如幻灯片文本和其他实体)关联视频。在本工作中,我们将介绍用于识别讲座实体的算法。具体来说,演讲视频中的实体是激光手势和指向手势以及幻灯片在视频中的位置及其文本和图像。我们的算法是在假设每个帧都知道幻灯片位置(单应性)并扩展了场景知识的前提下进行的。具体来说,手势可以通知幻灯片上何时何地发生了重要事件。我们还将展示对这些实体的识别如何帮助更好地理解演讲并在移动设备上实现节能。我们进行了一项用户研究,结果表明,基于幻灯片上的激光手势放大文本有助于将观看者的注意力吸引到相关文本上。我们还对真实的手机进行了实证测量,以确认选择性地调暗视频帧的不相关区域会显着降低能耗。

著录项

  • 作者

    Tung, Qiyam Junn.;

  • 作者单位

    The University of Arizona.;

  • 授予单位 The University of Arizona.;
  • 学科 Computer Science.
  • 学位 Ph.D.
  • 年度 2015
  • 页码 85 p.
  • 总页数 85
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号