...
首页> 外文期刊>JMLR: Workshop and Conference Proceedings >Approximation Schemes for ReLU Regression
【24h】

Approximation Schemes for ReLU Regression

机译:Relu回归的近似方案

获取原文
           

摘要

We consider the fundamental problem of ReLU regression, where the goal is to output the best fitting ReLU with respect to square loss given access to draws from some unknown distribution. We give the first efficient, constant-factor approximation algorithm for this problem assuming the underlying distribution satisfies some weak concentration and anti-concentration conditions (and includes, for example, all log-concave distributions). This solves the main open problem of Goel et al., who proved hardness results for any exact algorithm for ReLU regression (up to an additive $epsilon$). Using more sophisticated techniques, we can improve our results and obtain a polynomial-time approximation scheme for any subgaussian distribution. Given the aforementioned hardness results, these guarantees can not be substantially improved. Our main insight is a new characterization of {em surrogate losses} for nonconvex activations. While prior work had established the existence of convex surrogates for monotone activations, we show that properties of the underlying distribution actually induce strong convexity for the loss, allowing us to relate the global minimum to the activation’s {em Chow parameters}.
机译:我们考虑Relu回归的根本问题,其中目标是在给出的广场丢失上输出最佳拟合Relu,从而从一些未知的分发绘制。假设底层分布满足一些弱浓度和抗浓缩条件(并且包括所有对数凹发行分布),我们给出了这个问题的第一个有效的恒因子近似算法。这解决了Goel等人的主要开放问题。,据证明了对Relu回归的任何确切算法的硬度结果(最多为Addive $ epsilon $)。使用更复杂的技术,我们可以提高我们的结果,并获得任何子静脉分布的多项式近似方案。鉴于上述硬度结果,这些保证不能显着改善。我们的主要洞察力是非渗透激活的{ EM代理损失}的新表征。虽然事先工作已经建立了单调激活的凸代替代品的存在,但我们表明底层分布的属性实际上诱导了损失的强大凸,允许我们将全局最小值与激活的{ EM Chow参数}联系起来。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号