首页> 外文OA文献 >Learning Aligned Cross-Modal Representations from Weakly Aligned Data
【2h】

Learning Aligned Cross-Modal Representations from Weakly Aligned Data

机译:从弱对齐数据学习对齐的跨模态表示

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

People can recognize scenes across many different modalities beyond natural images. In this paper, we investigate how to learn cross-modal scene representations that transfer across modalities. To study this problem, we introduce a new cross-modal scene dataset. While convolutional neural networks can categorize cross-modal scenes well, they also learn an intermediate representation not aligned across modalities, which is undesirable for crossmodal transfer applications. We present methods to regularize cross-modal convolutional neural networks so that they have a shared representation that is agnostic of the modality. Our experiments suggest that our scene representation can help transfer representations across modalities for retrieval. Moreover, our visualizations suggest that units emerge in the shared representation that tend to activate on consistent concepts independently of the modality.
机译:人们可以识别自然图像以外的多种形式的场景。在本文中,我们研究了如何学习跨模式传递的跨模式场景表示。为了研究这个问题,我们引入了一个新的交叉模式场景数据集。尽管卷积神经网络可以很好地对交叉模式场景进行分类,但它们还学习了一种不跨模式对齐的中间表示形式,这对于交叉模式传输应用是不希望的。我们提出了规范化跨模态卷积神经网络的方法,以便它们具有与模态无关的共享表示。我们的实验表明,我们的场景表示可以帮助跨各种形式传输表示以进行检索。此外,我们的可视化结果表明,在共享表示形式中出现的单元倾向于在一致的概念上激活,而与模式无关。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号