首页> 外文会议>IEEE Conference on Computer Vision and Pattern Recognition >Jointly Modeling Embedding and Translation to Bridge Video and Language
【24h】

Jointly Modeling Embedding and Translation to Bridge Video and Language

机译:联合建模嵌入和翻译桥梁视频和语言

获取原文

摘要

Automatically describing video content with natural language is a fundamental challenge of computer vision. Recurrent Neural Networks (RNNs), which models sequence dynamics, has attracted increasing attention on visual interpretation. However, most existing approaches generate a word locally with the given previous words and the visual content, while the relationship between sentence semantics and visual content is not holistically exploited. As a result, the generated sentences may be contextually correct but the semantics (e.g., subjects, verbs or objects) are not true. This paper presents a novel unified framework, named Long Short-Term Memory with visual-semantic Embedding (LSTM-E), which can simultaneously explore the learning of LSTM and visual-semantic embedding. The former aims to locally maximize the probability of generating the next word given previous words and visual content, while the latter is to create a visual-semantic embedding space for enforcing the relationship between the semantics of the entire sentence and visual content. The experiments on YouTube2Text dataset show that our proposed LSTM-E achieves to-date the best published performance in generating natural sentences: 45.3% and 31.0% in terms of BLEU@4 and METEOR, respectively. Superior performances are also reported on two movie description datasets (M-VAD and MPII-MD). In addition, we demonstrate that LSTM-E outperforms several state-of-the-art techniques in predicting Subject-Verb-Object (SVO) triplets.
机译:自动描述具有自然语言的视频内容是计算机愿景的根本挑战。模型序列动态的经常性神经网络(RNN)吸引了对视觉解释的越来越关注。但是,大多数现有方法用给定的先前单词和视觉内容在本地生成一句话,而句子语义与视觉内容之间的关系不是全能地利用。结果,所生成的句子可以是上下文上的,而是语义(例如,受试者,动词或对象)不是真的。本文提出了一种新颖的统一框架,名为L长的短期内存,具有视觉语义嵌入(LSTM-E),可以同时探索LSTM和视觉语义嵌入的学习。前者旨在局部地最大化生成前一词和视觉内容的下一个单词的概率,而后者是创建用于执行整个句子和视觉内容的语义之间的关系的可视语义嵌入空间。 YouTube2Text数据集的实验表明,我们提出的LSTM-E分别在生成自然句子中实现了最佳公布的绩效:45.3%和31.0%,分别在Bleu @ 4和流星方面。在两部电影描述数据集(M-VAD和MPII-MD)上还报告了优越的性能。此外,我们证明LSTM-E在预测主语动词 - 对象(SVO)三元组时优于几种最先进的技术。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号