首页> 外文期刊>JMLR: Workshop and Conference Proceedings >Logarithmic Regret for Online Gradient Descent Beyond Strong Convexity
【24h】

Logarithmic Regret for Online Gradient Descent Beyond Strong Convexity

机译:对数后悔在线梯度下降超过强凸性

获取原文
           

摘要

Hoffman’s classical result gives a bound on the distance of a point from a convex and compact polytope in terms of the magnitude of violation of the constraints. Recently, several results showed that Hoffman’s bound can be used to derive strongly-convex-like rates for first-order methods for extit{offline} convex optimization of curved, though not strongly convex, functions, over polyhedral sets. In this work, we use this classical result for the first time to obtain faster rates for extit{online convex optimization} over polyhedral sets with curved convex, though not strongly convex, loss functions. We show that under several reasonable assumptions on the data, the standard extit{Online Gradient Descent} algorithm guarantees logarithmic regret. To the best of our knowledge, the only previous algorithm to achieve logarithmic regret in the considered settings is the extit{Online Newton Step} algorithm which requires quadratic (in the dimension) memory and at least quadratic runtime per iteration, which greatly limits its applicability to large-scale problems. In particular, our results hold for extit{semi-adversarial} settings in which the data is a combination of an arbitrary (adversarial) sequence and a stochastic sequence, which might provide reasonable approximation for many real-world sequences, or under a natural assumption that the data is low-rank. We demonstrate via experiments that the regret of OGD is indeed comparable to that of ONS (and even far better) on curved though not strongly-convex losses.
机译:霍夫曼(Hoffman)的经典结果以违反约束的程度来界定点到凸且紧凑的多面体的距离。最近,一些结果表明,霍夫曼边界可用于为多面体集上的曲线(虽然不是强凸)函数的 textit {offline}凸优化的一阶方法推导类似强凸的速率。在这项工作中,我们首次使用这种经典结果来获得具有弯曲凸(虽然不是强凸)损失函数的多面体集的 textit {在线凸优化}的更快速率。我们表明,在对数据的几个合理假设下,标准 textit {在线梯度下降}算法可确保对数后悔。据我们所知,在考虑的设置中唯一能够实现对数后悔的先前算法是 textit {在线牛顿步长}算法,该算法需要二次(维数)存储,并且每次迭代至少需要二次运行时间,这极大地限制了其适用于大规模问题。特别是,我们的结果适用于 textit {semi-adversarial}设置,其中数据是任意(对抗性)序列和随机序列的组合,可以为许多现实世界的序列或自然序列提供合理的近似值假设数据是低等级的。通过实验,我们证明了OGD的遗憾确实可以与ONS的遗憾相媲美(甚至更好),尽管曲线损失不大,但弯曲程度却不高。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号