首页> 外文会议>International Conference on Machine Learning, Optimization, and Data Science >Ensemble Kalman Filter Optimizing Deep Neural Networks: An Alternative Approach to Non-performing Gradient Descent
【24h】

Ensemble Kalman Filter Optimizing Deep Neural Networks: An Alternative Approach to Non-performing Gradient Descent

机译:Ensemble Kalman滤波器优化深神经网络:非执行梯度下降的替代方法

获取原文

摘要

The successful training of deep neural networks is dependent on initialization schemes and choice of activation functions. Non-optimally chosen parameter settings lead to the known problem of exploding or vanishing gradients. This issue occurs when gradient descent and backpropagation are applied. For this setting the Ensemble Kalman Filter (EnKF) can be used as an alternative optimizer when training neural networks. The EnKF does not require the explicit calculation of gradients or adjoints and we show this resolves the exploding and vanishing gradient problem. We analyze different parameter initializations, propose a dynamic change in ensembles and compare results to established methods.
机译:深度神经网络的成功培训取决于初始化方案和激活功能的选择。 非最佳选择的参数设置导致爆炸或消失渐变的已知问题。 应用梯度血统和逆产后发生此问题。 对于此设置,Ensemble Kalman滤波器(ENKF)可在培训神经网络时用作替代优化器。 ENKF不需要明确计算渐变或伴侣,我们展示了解决爆炸和消失的梯度问题。 我们分析了不同的参数初始化,提出了集合的动态变化,并将结果与已建立的方法进行比较。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号