首页> 外文期刊>International Journal of Parallel, Emergent and Distributed Systems >Multimedia enriched ontologies for video digital libraries
【24h】

Multimedia enriched ontologies for video digital libraries

机译:多媒体丰富的视频数字图书馆本体

获取原文
获取原文并翻译 | 示例

摘要

The development of appropriate tools and solutions to support effective access to video content is one of the main challenges for video digital libraries. Different techniques for manual and automatic annotation and retrieval have been proposed in recent years. It is a common practice to use linguistic ontologies for video annotation and retrieval: video elements are classified by establishing relationships between video contents and linguistic terms that identify domain concepts at different abstraction levels. However, although linguistic terms are appropriate to distinguish event and object categories, they are inadequate when they must describe specific or complex patterns of events or video entities. Instead, in these cases, pattern specifications can be better expressed using visual prototypes, either images or video clips, that capture the essence of the event or entity. High level concepts, expressed trough linguistic terms, and patterns specification, represented by visual prototypes, can be both organized into new extended ontologies where images or video clips are added to the ontologies as specification of linguistic terms. This paper presents algorithms and techniques that employ enriched ontologies for video annotation and retrieval, and discusses a solution for their implementation for the soccer video domain. An unsupervised clustering method is proposed in order to create multimedia enriched ontologies by defining visual prototypes that represent specific patterns of highlights and adding them as visual concepts to the ontology.An algorithm that uses multimedia enriched ontologies to perform automatic soccer video annotation is proposed and results for typical highlights are presented. Annotation is performed associating occurrences of events, or entities, to higher level concepts by checking their similarity to visual concepts that are hierarchically linked to higher level semantics, using a dynamic programming approach. Usage of reasoning on the ontology is shown, to create complex queries that comprise visual prototypes of actions, their temporal evolution and relations.
机译:开发适当的工具和解决方案以支持对视频内容的有效访问是视频数字图书馆的主要挑战之一。近年来,已经提出了用于手动和自动注释和检索的不同技术。使用语言本体进行视频注释和检索是一种常见的做法:通过在视频内容和语言术语之间建立关系来对视频元素进行分类,这些语言术语标识了不同抽象级别的域概念。但是,尽管语言术语适用于区分事件和对象类别,但是当它们必须描述事件或视频实体的特定或复杂模式时,它们并不足够。相反,在这些情况下,可以使用捕获事件或实体本质的视觉原型(图像或视频剪辑)更好地表达模式规范。可视语言原型表示的高级概念(表示槽语言术语)和模式规范都可以组织成新的扩展本体,其中将图像或视频剪辑作为语言术语的规范添加到本体中。本文介绍了利用丰富的本体进行视频注释和检索的算法和技术,并讨论了在足球视频领域中实现它们的解决方案。为了定义丰富的本体,提出了一种无监督的聚类方法,方法是定义代表亮点的特定模式的视觉原型并将其作为视觉概念添加到本体中。提出了一种使用丰富的本体执行自动足球视频注释的算法,并得出了结果介绍了典型的亮点。通过使用动态编程方法,通过检查事件或实体的出现与层次结构链接到高级语义的视觉概念的相似性,来将事件或实体的出现与高级概念相关联,从而执行注释。显示了对本体进行推理的用法,以创建复杂的查询,这些查询包含动作的视觉原型,其时间演变和关系。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号