首页> 外文会议>IEEE international conference on signal processing systems >Assessment and comparison of two pre-training methods for regularizing deep architecture neural networks in face recognition
【24h】

Assessment and comparison of two pre-training methods for regularizing deep architecture neural networks in face recognition

机译:评估人脸识别中正则化深度神经网络的两种预训练方法的评估和比较

获取原文

摘要

Training deep architecture neural networks have been considered widely recently. Besides the common and direct approach of training neural networks in which gradient-based training is used after choosing the appropriate structure, new approaches based on pre-training are more efficient in learning highly-varying complex functions. Not only gradient-based training is very slow, but also due to starting from random initialization, they generally get stuck in apparent local minimum and as the architecture gets deeper, it becomes more difficult to obtain good generalization. One way to solve this problem is to use pre-training and the multi-layer network is broken into some learning machines with the purpose of finding a good starting point in the function space of network. In this paper two pre-training methods are applied to obtain a suitable region of initial weights which is better than random initialization. Both two methods are implemented and used in a face recognition task on AT&T database. Results show that although first pretraining method is faster, the second method leads to better generalization and higher recognition rate.
机译:训练深度架构神经网络最近已被广泛考虑。除了在选择合适的结构后使用基于梯度的训练的神经网络训练的直接方法之外,基于预训练的新方法在学习高度变化的复杂函数方面更为有效。不仅基于梯度的训练非常慢,而且由于从随机初始化开始,它们通常陷入明显的局部最小值,并且随着体系结构的深入,获得良好的概括变得越来越困难。解决此问题的一种方法是使用预训练,并将多层网络分为一些学习机,目的是在网络的功能空间中找到一个良好的起点。本文采用两种预训练方法来获得合适的初始权重区域,该区域优于随机初始化。两种方法都在AT&T数据库上的人脸识别任务中实现并使用。结果表明,虽然第一种预训练方法速度更快,但是第二种方法却能更好地推广和提高识别率。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号