首页> 外文期刊>IEEE Transactions on Vehicular Technology >Common Semantic Representation Method Based on Object Attention and Adversarial Learning for Cross-Modal Data in IoV
【24h】

Common Semantic Representation Method Based on Object Attention and Adversarial Learning for Cross-Modal Data in IoV

机译:IoV中基于对象注意力和对抗学习的通用语义表示方法

获取原文
获取原文并翻译 | 示例
           

摘要

With the significant development of the Internet of Vehicles (IoV), various modal data, such as image and text, are emerging, which provide data support for good vehicle networking services. In order to make full use of the cross-modal data, we need to establish a common semantic representation to achieve effective measurement and comparison of different modal data. However, due to the heterogeneous distributions of cross-modal data, there exists a semantic gap between them. Although some deep neural network (DNN) based methods have been proposed to deal with this problem, there still exist several challenges: the qualities of the modality-specific features, the structure of the DNN, and the components of the loss function. In this paper, for representing cross-modal data in IoV, we propose a common semantic representation method based on object attention and adversarial learning (OAAL). To acquire high-quality modality-specific feature, in OAAL, we design an object attention mechanism, which links the cross-modal features effectively. To further alleviate the heterogeneous semantic gap, we construct a cross-modal generative adversarial network, which contains two parts: a generative model and a discriminative model. Besides, we also design a comprehensive loss function for the generative model to produce high-quality features. With a minimax game between the two models, we can construct a shared semantic space and generate the unified representations for cross-modal data. Finally, we apply our OAAL on retrieval task, and the results of the experiments have verified its effectiveness.
机译:随着车辆互联网(IoV)的飞速发展,各种模态数据(例如图像和文本)应运而生,这些数据为良好的车辆联网服务提供了数据支持。为了充分利用交叉模态数据,我们需要建立一个通用的语义表示,以实现对不同模态数据的有效度量和比较。但是,由于交叉模式数据的异构分布,它们之间存在语义鸿沟。尽管已提出了一些基于深度神经网络(DNN)的方法来解决此问题,但仍然存在一些挑战:特定于模态的特征的质量,DNN的结构以及损失函数的组成部分。在本文中,为了表示IoV中的跨模式数据,我们提出了一种基于对象注意和对抗学习(OAAL)的通用语义表示方法。为了获得高质量的特定于情态的功能,在OAAL中,我们设计了一种对象关注机制,该机制有效地链接了跨模式的功能。为了进一步缓解异构语义鸿沟,我们构建了一个跨模态的生成对抗网络,该网络包含两个部分:生成模型和判别模型。此外,我们还为生成模型设计了一个综合损失函数,以产生高质量的特征。通过两个模型之间的极大极小游戏,我们可以构造一个共享的语义空间并生成跨模式数据的统一表示。最后,我们将OAAL应用于检索任务,实验结果证明了其有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号