...
首页> 外文期刊>ACM transactions on multimedia computing communications and applications >Video Decolorization Based on the CNN and LSTM Neural Network
【24h】

Video Decolorization Based on the CNN and LSTM Neural Network

机译:基于CNN和LSTM神经网络的视频脱色

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

Video decolorization is the process of transferring three-channel color videos into single-channel grayscale videos, which is essentially the decolorization operation of video frames. Most existing video decolorization algorithms directly apply image decolorization methods to decolorize video frames. However, if we only take the single-frame decolorization result into account, it will inevitably cause temporal inconsistency and flicker phenomenon meaning that the same local content between continuous video frames may display different gray values. In addition, there are often similar local content features between video frames, which indicates redundant information. To solve the preceding problems, this article proposes a novel video decolorization algorithm based on the convolutional neural network and the long short-term memory neural network. First, we design a local semantic content encoder to learn and extract the same local content of continuous video frames, which can better preserve the contrast of video frames. Second, a temporal feature controller based on the bi-directional recurrent neural networks with Long short-term memory units is employed to refine the local semantic features, which can greatly maintain temporal consistency of the video sequence to eliminate the flicker phenomenon. Finally, we take advantages of deconvolution to decode the features to produce the grayscale video sequence. Experiments have indicated that our method can better preserve the local contrast of video frames and the temporal consistency over the state of the-art.
机译:视频脱色是将三声道彩色视频传输到单通道灰度视频的过程,这基本上是视频帧的偏差操作。大多数现有的视频偏差算法直接应用图像脱色方法以脱色视频帧。但是,如果我们只考虑单帧脱色结果,则它将不可避免地引起时间不一致和闪烁现象,这意味着连续视频帧之间的相同本地内容可以显示不同的灰度值。此外,视频帧之间通常存在类似的本地内容特征,其指示冗余信息。为了解决前面的问题,本文提出了一种基于卷积神经网络和长短期记忆神经网络的新型视频脱色算法。首先,我们设计一个本地语义内容编码器来学习和提取连续视频帧的相同本地内容,这可以更好地保留视频帧的对比度。其次,采用基于具有长短期存储器单元的双向复发性神经网络的时间特征控制器来优化局部语义特征,这可以大大维持视频序列的时间一致性以消除闪烁现象。最后,我们采取了解压缩的优势来解码功能以产生灰度视频序列。实验表明,我们的方法可以更好地保留视频帧的局部对比度和最先进的局部一致性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号