首页> 外文OA文献 >A full-scale semantic content-based model for interactive multimedia information systems.
【2h】

A full-scale semantic content-based model for interactive multimedia information systems.

机译:一种基于全尺寸语义内容的交互式多媒体信息系统模型。

摘要

Issues of syntax have dominated research in multimedia information systems (MMISs), with video developing as a technology of images and audio as one of signals. But when we use video and audio, we do so for their content. This is a semantic issue. Current research in multimedia on semantic content-based models has adopted a structure-oriented approach, where video and audio content is described on a frame-by-frame or segment-by-segment basis (where a segment is an arbitrary set of contiguous frames). This approach has failed to cater for semantic aspects, and thus has not been fully effective when used within an MMIS. The research undertaken for this thesis reveals seven semantic aspects of video and audio: (1) explicit media structure; (2) objects; (3) spatial relationships between objects; (4) events and actions involving objects; (5) temporal relationships between events and actions; (6) integration of syntactic and semantic information; and (7) direct user-media interaction. This thesis develops a full-scale semantic content-based model that caters for the above seven semantic aspects of video and audio. To achieve this, it uses an entities of interest approach, instead of a structure-oriented one, where the MMIS integrates relevant semantic content-based information about video and audio with information about the entities of interest to the system, e.g. mountains, vehicles, employees. A method for developing an interactive MMIS that encompasses the model is also described. Both the method and the model are used in the development of ARISTOTLE, an interactive instructional MMIS for teaching young children about zoology, in order to demonstrate their operation.
机译:语法问题已成为多媒体信息系统(MMIS)研究的主导,视频技术是图像技术,音频技术是信号技术之一。但是,当我们使用视频和音频时,我们是针对它们的内容。这是一个语义问题。多媒体在基于语义内容的模型上的当前研究已采用了一种面向结构的方法,其中视频和音频内容是在逐帧或逐段的基础上进行描述的(其中,分段是任意连续帧的集合) )。这种方法无法满足语义方面的要求,因此在MMIS中使用时并没有完全有效。本文的研究揭示了视频和音频的七个语义方面:(1)明确的媒体结构; (2)物体; (3)物体之间的空间关系; (四)涉及对象的事件和动作; (5)事件与动作之间的时间关系; (6)句法和语义信息的整合; (7)直接的用户与媒体互动。本文开发了一种基于语义内容的全方位模型,该模型可以满足视频和音频的上述七个语义方面的需求。为了实现这一点,它使用感兴趣实体方法,而不是面向结构的方法,在该方法中,MMIS将有关视频和音频的基于语义内容的相关信息与有关系统感兴趣的实体的信息集成在一起,例如:山脉,车辆,员工。还描述了一种开发包含模型的交互式MMIS的方法。该方法和模型都用于ARISTOTLE的开发中。ARISTOTLE是一种交互式的教学MMIS,用于向幼儿讲授动物学,以演示其操作方法。

著录项

  • 作者

    Agius Harry Wayne;

  • 作者单位
  • 年度 1997
  • 总页数
  • 原文格式 PDF
  • 正文语种 en
  • 中图分类

相似文献

  • 外文文献
  • 中文文献
  • 专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号