首页> 外文OA文献 >Video Summarization With Attention-Based Encoder–Decoder Networks
【2h】

Video Summarization With Attention-Based Encoder–Decoder Networks

机译:基于关注的编码器解码器网络的视频概述

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

This paper addresses the problem of supervised video summarization byformulating it as a sequence-to-sequence learning problem, where the input is asequence of original video frames, the output is a keyshot sequence. Our keyidea is to learn a deep summarization network with attention mechanism to mimicthe way of selecting the keyshots of human. To this end, we propose a novelvideo summarization framework named Attentive encoder-decoder networks forVideo Summarization (AVS), in which the encoder uses a Bidirectional LongShort-Term Memory (BiLSTM) to encode the contextual information among the inputvideo frames. As for the decoder, two attention-based LSTM networks areexplored by using additive and multiplicative objective functions,respectively. Extensive experiments are conducted on three video summarizationbenchmark datasets, i.e., SumMe, TVSum, and YouTube. The results demonstratethe superiority of the proposed AVS-based approaches against thestate-of-the-art approaches, with remarkable improvements from 3% to 11% on thethree datasets, respectively.
机译:本文解决了监督视频摘要的问题,通过将其作为序列到序列学习问题,其中输入是原始视频帧的序列,输出是键序列。我们的keyidea是通过注意机制学习深度摘要网络,以模仿选择人类钥匙的方式。为此,我们提出了一种名为注意力编码器 - 解码器网络的新颖性摘要框架,其概述(AVS),其中编码器使用双向延长术语存储器(BILSTM)来对InputVideo帧之间的上下文信息进行编码。至于解码器,通过使用添加剂和乘法物理函数来绘制两个基于关注的LSTM网络。广泛的实验是在三个视频摘要中进行的,即夏季,TVSUM和YouTube进行。结果证明了拟议的基于AVS的方法与最近的方法的方法,分别从3%到11%的替代数据集中的显着改善。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号