首页> 外文会议>International Conference on Machine Learning >Double Trouble in Double Descent: Bias and Variance(s) in the Lazy Regime
【24h】

Double Trouble in Double Descent: Bias and Variance(s) in the Lazy Regime

机译:双重血统中的双重麻烦:懒惰政权中的偏见和差异

获取原文

摘要

Deep neural networks can achieve remarkable generalization performances while interpolating the training data; rather than the U-curve emblematic of the bias-variance trade-off, their test error often follows a "double descent" curve - a mark of the beneficial role of overparametrization. In this work, we develop a quantitative theory for this phenomenon in the context of highdimensional random features regression. We obtain a precise asymptotic expression for the biasvariance decomposition of the test error, and show that the bias displays a phase transition at the interpolation threshold, beyond it which it remains constant. We disentangle the variances stemming from the sampling of the dataset, from the additive noise corrupting the labels, and from the initialization of the weights. Following up on (Geiger et al., 2019a), we demonstrate that the latter two contributions are the crux of the double descent: they lead to the overfitting peak at the interpolation threshold and to the decay of the test error upon overparametrization. We quantify how they are suppressed by averaging the outputs of independently initialized estimators, and compare this ensembling procedure with overparametrization and regularization. Finally, we present numerical experiments on a standard deep learning setup to show that our results are relevant to the lazy regime of deep neural networks.
机译:深度神经网络在对训练数据进行插值时,可以获得显著的泛化性能;与代表偏差-方差权衡的U型曲线不同,他们的测试误差通常遵循“双下降”曲线——这标志着过度参数化的有利作用。在这项工作中,我们在高维随机特征回归的背景下为这一现象发展了一个定量理论。我们得到了测试误差的偏差方差分解的精确渐近表达式,并表明偏差在插值阈值处显示相变,超过该阈值时,偏差保持不变。我们从数据集采样、破坏标签的加性噪声和权重初始化中分离出方差。继(Geiger等人,2019a)之后,我们证明后两种贡献是双重下降的关键:它们导致插值阈值处的过度拟合峰值,以及过度参数化后测试误差的衰减。我们通过对独立初始化的估计器的输出求平均来量化它们是如何被抑制的,并将这种置乱过程与过参数化和正则化进行比较。最后,我们在一个标准的深度学习装置上进行了数值实验,以表明我们的结果与深度神经网络的惰性状态有关。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号