首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >A Model of Double Descent for High-Dimensional Logistic Regression
【24h】

A Model of Double Descent for High-Dimensional Logistic Regression

机译:高维Logistic回归的双下降模型

获取原文

摘要

We consider a model for logistic regression where only a subset of features of size p is used for training a linear classifier over n training samples. The classifier is obtained by running gradient-descent (GD) on the logistic-loss. For this model, we investigate the dependence of the classification error on the overparameterization ratio κ = p. First, building on known deterministic results on convergence properties of the GD, we uncover a phase-transition phenomenon for the case of Gaussian features: the classification error of GD is the same as that of the maximum-likelihood (ML) solution when κ < κ, and that of the max-margin (SVM) solution when κ > κ. Next, using the convex Gaussian min-max theorem (CGMT), we sharply characterize the performance of both the ML and SVM solutions. Combining these results, we obtain curves that explicitly characterize the test error of GD for varying values of κ. The numerical results validate the theoretical predictions and unveil "double-descent" phenomena that complement similar recent observations in linear regression settings.
机译:我们考虑用于逻辑回归的模型,其中仅大小为p的特征子集用于训练n个训练样本上的线性分类器。通过对逻辑损失运行梯度下降(GD)获得分类器。对于此模型,我们研究了分类误差对过参数化比率κ= p / n的依赖性。首先,基于已知的GD收敛性的确定性结果,我们发现了高斯特征情况下的相变现象:当κ<时,GD的分类误差与最大似然(ML)解相同。 κ ,以及当κ>κ时的最大保证金(SVM)解的值 。接下来,使用凸高斯最小极大定理(CGMT),我们清晰地描述了ML和SVM解决方案的性能。结合这些结果,我们获得了曲线,这些曲线明确地表征了针对κ值变化的GD的测试误差。数值结果验证了理论预测并揭示了“双下降”现象,该现象补充了线性回归设置中类似的近期观察结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号