首页> 外文会议>AAAI Conference on Artificial Intelligence >Unicoder-VL: A Universal Encoder for Vision and Language by Cross-Modal Pre-Training
【24h】

Unicoder-VL: A Universal Encoder for Vision and Language by Cross-Modal Pre-Training

机译:Unicoder-VL:跨模型预培训的视觉和语言的通用编码器

获取原文

摘要

We propose Unicoder-VL, a universal encoder that aims to learn joint representations of vision and language in a pre-training manner. Borrow ideas from cross-lingual pre-trained models, such as XLM (Lample and Conneau 2019) and Unicoder (Huang et al. 2019), both visual and linguistic contents are fed into a multi-layer Transformer (Vaswani et al. 2017) for the cross-modal pre-training, where three pre-trained tasks are employed, including Masked Language Modeling (MLM), Masked Object Classification (MOC) and Visual-linguistic Matching (VLM). The first two tasks learn context-aware representations for input tokens based on linguistic and visual contents jointly. The last task tries to predict whether an image and a text describe each other. After pretraining on large-scale image-caption pairs, we transfer Unicoder-VL to caption-based image-text retrieval and visual commonsense reasoning, with just one additional output layer. We achieve state-of-the-art or comparable results on both two tasks and show the powerful ability of the cross-modal pre-training.
机译:我们提出Unicoder-VL,这是一个普遍的编码器,旨在以预先培训方式学习视觉和语言的联合表示。借用XLM(2019年Lopple和Compeau)和Unicoder(Huang等,2019)中的跨语预训练模型的想法,将视觉和语言内容融入多层变压器(Vaswani等,2017)对于跨模型预培训,其中采用了三个预先训练的任务,包括屏蔽语言建模(MLM),屏蔽对象分类(MOC)和视觉语言匹配(VLM)。前两个任务学习基于语言和视觉内容的输入令牌的上下文感知表示。最后一项任务试图预测图像和文本是否彼此描述。在预先绘制大规模图像标题对之后,我们将Unicoder-VL传输到基于标题的图像文本检索和可视致辞推理,只需一个额外的输出层。我们在两项任务中实现最先进的或可比结果,并显示出跨模型预培训的强大能力。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号