首页> 外文会议>AAAI Conference on Artificial Intelligence >A Sharper Generalization Bound for Divide-and-Conquer Ridge Regression
【24h】

A Sharper Generalization Bound for Divide-and-Conquer Ridge Regression

机译:一个较小的泛化界限,用于分割和征收山脊回归

获取原文

摘要

We study the distributed machine learning problem where the n feature-response pairs are partitioned among m machines uniformly at random. The goal is to approximately solve an empirical risk minimization (ERM) problem with the minimum amount of communication. The divide-and-conquer (DC) method, which was proposed several years ago, lets every worker machine independently solve the same ERM problem using its local feature-response pairs and the driver machine combine the solutions. This approach is in one-shot and thereby extremely communication-efficient. Although the DC method has been studied by many prior works, reasonable generalization bound has not been established before this work. For the ridge regression problem, we show that the prediction error of the DC method on unseen test samples is at most ε times larger than the optimal. There have been constant-factor bounds in the prior works, their sample complexities have a quadratic dependence on d, which does not match the setting of most real-world problems. In contrast, our bounds are much stronger. First, our 1 + ε error bound is much better than their constant-factor bounds. Second, our sample complexity is merely linear with d.
机译:我们研究了分布式机器学习问题,其中n个功能响应对在随机均匀地在M机器中划分。目标是大致解决了具有最小通信量的经验风险最小化(ERM)问题。几年前提议的分派和征管(DC)方法,让每个工人机器使用其本地特征响应对和驱动机器组合解决方案来独立解决相同的ERM问题。这种方法是一次性的,从而极其沟通效率。虽然已经通过许多事先作用研究了DC方法,但在这项工作之前尚未建立合理的泛化界定。对于Ridge回归问题,我们表明DC方法对看不见的试样上的预测误差是大部分大于最佳的ε倍。先前作品中存在恒定因子界限,它们的样本复杂性对D具有二次依赖性,这与大多数现实世界问题的设置不符。相比之下,我们的界限更强大。首先,我们的1 +ε错误绑定比其恒因因子界限好得多。其次,我们的样本复杂性仅用D线性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号