首页> 外文期刊>JMLR: Workshop and Conference Proceedings >A tight excess risk bound via a unified PAC-Bayesian–Rademacher–Shtarkov–MDL complexity
【24h】

A tight excess risk bound via a unified PAC-Bayesian–Rademacher–Shtarkov–MDL complexity

机译:通过统一的PAC-Bayesian–Rademacher–Shtarkov–MDL复杂性来约束紧密的过剩风险

获取原文
           

摘要

We present a novel notion of complexity that interpolates between and generalizes some classic complexity notions in learning theory: for empirical risk minimization (ERM) with arbitrary bounded loss, it is upper bounded in terms of data-independent Rademacher complexity; for generalized Bayesian estimators, it is upper bounded by the data-dependent information (KL) complexity. For ERM, the new complexity reduces to normalized maximum likelihood complexity, i.e., a minimax log-loss individual sequence regret. Our first main result bounds excess risk in terms of the new complexity. Our second main result links the new complexity to $L_2(P)$ entropy via Rademacher complexity, generalizing earlier results of Opper, Haussler, Lugosi, and Cesa-Bianchi who covered the log-loss case with $L_infty$ entropy. Together, these results recover optimal bounds for VC-type and large (polynomial entropy) classes, replacing local Rademacher complexities by a simpler analysis which almost completely separates the two aspects that determine the achievable rates: ‘easiness’ (Bernstein) conditions and model complexity.
机译:我们提出一种新颖的复杂性概念,在学习理论中进行插值并概括了一些经典的复杂性概念:对于具有任意有界损失的经验风险最小化(ERM),它在数据独立的Rademacher复杂性方面是上限。对于广义贝叶斯估计量,它受数据相关信息(KL)复杂度的上限。对于ERM,新的复杂度降低为归一化的最大似然复杂度,即最小最大对数损失单个序列表示遗憾。我们的第一个主要结果在新的复杂性方面限制了过多的风险。我们的第二个主要结果是通过Rademacher复杂度将新的复杂度与$ L_2(P)$熵联系起来,推广了Opper,Haussler,Lugosi和Cesa-Bianchi的早期结果,后者用$ L_ infty $熵覆盖了对数损失案例。综合起来,这些结果恢复了VC型和大型(多项式熵)类别的最佳界限,通过更简单的分析代替了局部Rademacher复杂度,该分析几乎完全分离了确定可实现速率的两个方面:“简便性”(Bernstein)条件和模型复杂性。

著录项

相似文献

  • 外文文献
  • 中文文献
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号