首页> 外文期刊>Neurocomputing >A fast and efficient pre-training method based on layer-by-layer maximum discrimination for deep neural networks
【24h】

A fast and efficient pre-training method based on layer-by-layer maximum discrimination for deep neural networks

机译:基于分层最大判别的深度神经网络快速高效的预训练方法

获取原文
获取原文并翻译 | 示例

摘要

In this paper, through extension of the present methods and based on error minimization, two fast and efficient layer-by-layer pre-training methods are proposed for initializing deep neural network (DNN) weights. Due to confrontation with a large number of local minima, DNN training often does not converge. By proper initializing of DNN weights instead of random values at the beginning of the training, it is possible to avoid many local minima. The first version of the proposed method is for pre-training the deep bottleneck neural network (DBNN) in which the DBNN is broken down to some corresponding single-hidden-layer bottleneck neural networks (BNN) which must be trained first. The weight values resulting from their training are then applied in the DBNN. The proposed method was utilized to pre-train a five-hidden-layer DBNN to extract the non-linear principal components of face images in the Bosphorus database. A comparison of the randomly initialized DBNN result with pre-trained DBNN by the layer-by-layer pre-training method shows that this method not only increased the convergence rate of training but also improved its generalizability. Furthermore, it has been shown that this method yields higher efficiency and convergence speed in comparison with some of the previous pre-training methods. This paper also presents the bidirectional version of the layer-by-layer pre-training method for hetero-associative DNN pre-training. This method pre-trains DNN weights in forward and backward manner in parallel. Bidirectional layer-by-layer pre-training was utilized to pre-train the classifier DNN weights, and revealed that both the training speed and the recognition rate were improved in Bosphorus and CK+ databases. (C) 2015 Elsevier B.V. All rights reserved.
机译:本文在扩展现有方法的基础上,基于误差最小化,提出了两种快速有效的逐层预训练方法来初始化深度神经网络权重。由于面对大量局部最小值,DNN培训通常无法收敛。通过在训练开始时正确初始化DNN权重而不是随机值,可以避免许多局部最小值。所提出方法的第一个版本是用于预训练深度瓶颈神经网络(DBNN),其中将DBNN分解为一些相应的单隐藏层瓶颈神经网络(BNN),必须首先对其进行训练。然后将其训练所得的权重值应用于DBNN。该方法被用于对五层DBNN进行预训练,以提取Bosphorus数据库中人脸图像的非线性主成分。通过逐层预训练方法对随机初始化的DBNN结果与预训练的DBNN的比较表明,该方法不仅提高了训练的收敛速度,而且提高了其推广性。此外,已经表明,与一些先前的预训练方法相比,该方法产生更高的效率和收敛速度。本文还提出了用于异构关联DNN预训练的逐层预训练方法的双向版本。该方法以并行方式向前和向后预训练DNN权重。利用双向逐层预训练对分类器DNN权重进行预训练,发现在Bosphorus和CK +数据库中,训练速度和识别率均得到了改善。 (C)2015 Elsevier B.V.保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号