首页> 外文期刊>应用数学与计算数学学报(英文) >Drop-Activation:Implicit Parameter Reduction and Harmonious Regularization
【24h】

Drop-Activation:Implicit Parameter Reduction and Harmonious Regularization

机译:Drop-Activation:Implicit Parameter Reduction and Harmonious Regularization

获取原文
获取原文并翻译 | 示例
       

摘要

Overfitting frequently occurs in deep learning.In this paper,we propose a novel regularization method called drop-activation to reduce overfitting and improve generalization.The key idea is to drop nonlinear activation functions by setting them to be identity functions randomly during training time.During testing,we use a deterministic network with a new activation function to encode the average effect of dropping activations randomly.Our theoretical analyses support the regularization effect of drop-activation as implicit parameter reduction and verify its capability to be used together with batch normalization(Iolfe and Szegedy in Batch normalization:accelerating deep network training by reducing internal covariate shift.arXiv:1502.03167,2015).The experimental results on CIFAR10,CIFAR100,SVHN,EMNIST,and ImageNet show that drop-activation generally improves the performance of popular neural network architectures for the image classification task.Furthermore,as a regularizer drop-activation can be used in harmony with standard training and regularization techniques such as batch normalization and AutoAugment(Cubuk et al.in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,pp.113-123,2019).The code is available at https://github.com/LeungSamWai/Drop-Activ ation.

著录项

获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号