首页> 外文期刊>JMLR: Workshop and Conference Proceedings >An Online Algorithm for Smoothed Regression and LQR Control
【24h】

An Online Algorithm for Smoothed Regression and LQR Control

机译:平滑回归和LQR控制的在线算法

获取原文
           

摘要

We consider Online Convex Optimization (OCO) in the setting where the costs are $m$-strongly convex and the online learner pays a switching cost for changing decisions between rounds. We show that the recently proposed Online Balanced Descent (OBD) algorithm is constant competitive in this setting, with competitive ratio $3 + O(1/m)$, irrespective of the ambient dimension. Additionally, we show that when the sequence of cost functions is $epsilon$-smooth, OBD has near-optimal dynamic regret and maintains strong per-round accuracy. We demonstrate the generality of our approach by showing that the OBD framework can be used to construct competitive algorithms for a variety of online problems across learning and control, including online variants of ridge regression, logistic regression, maximum likelihood estimation, and LQR control.
机译:我们考虑在线凸优化(OCO),在这种情况下成本为$ m $-凸,而在线学习者为轮次之间的决策变更支付转换成本。我们表明,最近提出的在线平衡下降(OBD)算法在这种情况下具有恒定的竞争能力,无论环境尺寸如何,竞争比为$ 3 + O(1 / m)$。此外,我们证明,当成本函数序列为ε平滑时,OBD具有近乎最佳的动态后悔效果,并保持了很高的每轮准确性。我们通过证明OBD框架可用于构建跨学习和控制的各种在线问题的竞争算法来证明我们方法的通用性,这些算法包括岭回归,逻辑回归,最大似然估计和LQR控制的在线变体。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号