首页> 外文期刊>Neural processing letters >Enhance the Performance of Deep Neural Networks via L2 Regularization on the Input of Activations
【24h】

Enhance the Performance of Deep Neural Networks via L2 Regularization on the Input of Activations

机译:通过L2正则化在激活输入中通过L2正常化提升深神经网络的性能

获取原文
获取原文并翻译 | 示例
           

摘要

Deep neural networks (DNNs) are witnessing increasing attention in machine learning. However, the information propagation is becoming increasingly difficult as the networks get deeper, which makes the optimization of DNN extremely hard. One reason of this difficulty is saturation of hidden units. In this paper, we propose a novel methodology named RegA to decrease the influences of saturation on ReLU-DNNs (DNNs with ReLU). Instead of changing the activation functions or the initialization strategy, our methodology explicitly encourage the pre-activation to be out of the saturation region. Specifically, we add an auxiliary objective induced by L2-norm of the pre-activation values to the optimization problem. The auxiliary objective could help to active more units and promote effective information propagation in ReLU-DNNs. By conducting experiments on several large-scale real datasets, we demonstrate better representations could be learned by using RegA and the method help ReLU-DNNs get better performance on convergence and accuracy.
机译:深度神经网络(DNN)正在目睹机器学习中的注意力越来越多。然而,随着网络更深的,信息传播变得越来越困难,这使得DNN非常难以优化。这种困难的一个原因是隐藏单元的饱和​​。在本文中,我们提出了一种名为Rega的新型方法,以减少饱和对Relu-DNN的影响(DNN与Relu)。我们的方法没有改变激活功能或初始化策略,而不是改变激活功能或初始化策略,明确鼓励预先激活超出饱和区域。具体地,我们将由预激活值的L2-Norm的辅助物镜添加到优化问题。辅助目标可以有助于激活更多单元并促进Relu-DNN中的有效信息传播。通过对几个大型实际数据集进行实验,我们通过使用Rega来了解更好的表示,并且该方法有助于Relu-DNN在收敛和准确性上获得更好的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号