首页> 外文期刊>Multimedia Tools and Applications >A novel weight initialization with adaptive hyper-parameters for deep semantic segmentation
【24h】

A novel weight initialization with adaptive hyper-parameters for deep semantic segmentation

机译:深度语义分割具有自适应超参数的新型重量初始化

获取原文
获取原文并翻译 | 示例
           

摘要

The semantic segmentation process divides an image into its constituent objects and background by assigning a corresponding class label to each pixel in the image. Semantic segmentation is an important area in computer vision with wide practical applications. The contemporary semantic segmentation approaches are primarily based on two types of deep neural networks architectures i.e., symmetric and asymmetric networks. Both types of networks consist of several layers of neurons which are arranged in two sections called encoder and decoder. The encoder section receives the input image and the decoder section outputs the segmented image. However, both sections in symmetric networks have the same number of layers and the number of neurons in an encoder layer is the same as that of the corresponding layer in the decoder section but asymmetric networks do not strictly follow such one-one correspondence between encoder and decoder layers. At the moment, SegNet and ESNet are the two leading state-of-the-art symmetric encoder-decoder deep neural network architectures. However, both architectures require extensive training for good generalization and need several hundred epochs for convergence. This paper aims to improve the convergence and enhance network generalization by introducing two novelties into the network training process. The first novelty is a weight initialization method and the second contribution is an adaptive mechanism for dynamic layer learning rate adjustment in training loop. The proposed initialization technique uses transfer learning to initialize the encoder section of the network, but for initialization of decoder section, the weights of the encoder section layers are copied to the corresponding layers of the decoder section. The second contribution of the paper is an adaptive layer learning rate method, wherein the learning rates of the encoder layers are updated based on a metric representing the difference between the probability distributions of the input images and encoder weights. Likewise, the learning rates of the decoder layers are updated based on the difference between the probability distributions of the output labels and decoder weights. Intensive empirical validation of the proposed approach shows significant improvement in terms of faster convergence and generalization.
机译:语义分割过程通过将相应的类标签分配给图像中的每个像素来将图像划分为其组成对象和背景。语义分割是具有广泛实际应用的计算机视觉中的重要领域。当代语义分割方法主要基于两种类型的深神经网络架构I.E.,对称和非对称网络。这两种类型的网络由几层神经元组成,该神经元由称为编码器和解码器的两个部分。编码器部分接收输入图像,解码器部分输出分段图像。然而,对称网络中的两个部分具有相同数量的层数,编码器层中的神经元数与解码器部分中的相应层的数量相同,但是不对称网络不会严格遵循编码器之间的这种一体对应关系解码器层。目前,SEGNET和ESNET是两个领先的最先进的对称编码器解码器深神经网络架构。然而,两种架构都需要广泛的培训良好的概括,需要几百个时期的收敛性。本文旨在通过将两份Noveltize引入网络培训过程来提高收敛性和提高网络泛化。第一新颖性是重量初始化方法,第二贡献是用于训练循环中的动态层学习速率调整的自适应机制。所提出的初始化技术使用转移学习来初始化网络的编码器部分,而是用于解码器部分的初始化,编码器部分层的权重被复制到解码器部分的相应层。纸张的第二贡献是自适应层学习率方法,其中基于表示输入图像和编码器权重之间的差异之间的差异的度量来更新编码器层的学习速率。同样,基于输出标签和解码器权重之间的概率分布之间的差异来更新解码器层的学习速率。提出的方法的密集实证验证表现出更快的收敛性和泛化方面的显着改善。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号