【24h】

Understanding Dropout

机译:了解辍学

获取原文
获取外文期刊封面目录资料

摘要

Dropout is a relatively new algorithm for training neural networks which relies on stochastically "dropping out" neurons during training in order to avoid the co-adaptation of feature detectors. We introduce a general formalism for studying dropout on either units or connections, with arbitrary probability values, and use it to analyze the averaging and regularizing properties of dropout in both linear and non-linear networks. For deep neural networks, the averaging properties of dropout are characterized by three recursive equations, including the approximation of expectations by normalized weighted geometric means. We provide estimates and bounds for these approximations and corroborate the results with simulations. Among other results, we also show how dropout performs stochastic gradient descent on a regularized error function.
机译:辍学是训练神经网络的相对较新的算法,其在训练期间依赖于随机的“掉落”神经元,以避免特征探测器的共同适应。我们介绍了一种用于研究任意概率值的单位或连接的丢失的一般形式主义,并使用它来分析线性和非线性网络中丢失的平均和正规化。对于深神经网络,辍学的平均性质的特征在于三个递归方程,包括通过归一化加权几何手段的预期近似。我们为这些近似提供估计和界限,并通过模拟证实结果。除其他结果之外,我们还展示了如何在正常的错误功能上表现随机梯度下降。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号