首页> 外文会议>International Conference on Artificial Neural Networks >Multi-modal Associative Storage and Retrieval Using Hopfield Auto-associative Memory Network
【24h】

Multi-modal Associative Storage and Retrieval Using Hopfield Auto-associative Memory Network

机译:使用Hopfield自联想存储网络进行多模式联想存储和检索

获取原文

摘要

Recently we presented text storage and retrieval in an auto-associative memory framework using the Hopfield neural-network. This realized the ideal functionality of Hopfield network as a content-addressable information retrieval system. In this paper, we extend this result to multi-modal patterns, namely, images with text captions and show that the Hopfield network indeed can store and retrieve such multimodal patterns even in an auto-associative setting. Within this framework, we examine two central issues such as (i) performance characterization to show that the O(N) capacity of the Hopfield network for a network of size N neurons under the Pseudo-inverse learning rule is still retained in the multi-modal case, and (ii) the retrieval dynamics of the multi-modal pattern (i.e., image and caption together) under various types of queries such as image+caption, image only and caption only, in line with a typical multi-modal retrieval system where the entire multi-modal pattern is expected to be retrieved even with a partial query pattern from any of the modalities. We present results related to these two issues on a large database of 7000+ captioned-images and establish the practical scalability of both the storage capacity and the retrieval robustness of the Hopfield network for content-addressable retrieval of multi-modal patterns. We point to the potential of this work to extend to a more wider definition of multi-modality as in multi-media content, with various modalities such as video (image sequence) synchronized with sub-title text, speech, music and non-speech.
机译:最近,我们提出了使用Hopfield神经网络在自动联想存储框架中进行文本存储和检索的方法。这实现了Hopfield网络作为内容可寻址信息检索系统的理想功能。在本文中,我们将此结果扩展到多模式模式,即带有文本标题的图像,并表明Hopfield网络的确可以在自动关联的情况下存储和检索这种多模式模式。在此框架内,我们研究了两个主要问题,例如(i)性能表征,以表明在伪逆学习规则下,对于大小为N的神经元网络,Hopfield网络的O(N)容量仍然保留。模态案例,以及(ii)在各种类型的查询(例如,图像+字幕,仅图像和仅字幕)下的多模式模式(即图像和标题一起)的检索动态,这与典型的多模式检索相符该系统中,即使使用任何一种模式的部分查询模式,都希望可以检索到整个多模式模式。我们在具有7000多个带字幕图像的大型数据库上呈现与这两个问题相关的结果,并建立了针对多模式模式的内容寻址检索的Hopfield网络的存储容量和检索鲁棒性的实用可扩展性。我们指出这项工作有潜力扩展到多媒体内容中对多模式的更广泛定义,并具有各种模式,例如与字幕文本,语音,音乐和非语音同步的视频(图像序列) 。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号