首页> 外文期刊>Automatic Control, IEEE Transactions on >An Asynchronous Mini-Batch Algorithm for Regularized Stochastic Optimization
【24h】

An Asynchronous Mini-Batch Algorithm for Regularized Stochastic Optimization

机译:正则随机优化的异步小批量算法

获取原文
获取原文并翻译 | 示例
           

摘要

Mini-batch optimization has proven to be a powerful paradigm for large-scale learning. However, the state-of-the-art parallel mini-batch algorithms assume synchronous operation or cyclic update orders. When worker nodes are heterogeneous (due to different computational capabilities or different communication delays), synchronous and cyclic operations are inefficient since they will leave workers idle waiting for the slower nodes to complete their computations. In this paper, we propose an asynchronous mini-batch algorithm for regularized stochastic optimization problems with smooth loss functions that eliminates idle waiting and allows workers to run at their maximal update rates. We show that by suitably choosing the step-size values, the algorithm achieves a rate of the order O(1/T√) for general convex regularization functions, and the rate O(1/T) for strongly convex regularization functions, where T is the number of iterations. In both cases, the impact of asynchrony on the convergence rate of our algorithm is asymptotically negligible, and a near-linear speed-up in the number of workers can be expected. Theoretical results are confirmed in real implementations on a distributed computing infrastructure.
机译:小批量优化已被证明是大规模学习的强大范例。但是,最新的并行微型批处理算法假定同步操作或循环更新顺序。当工作程序节点异构(由于不同的计算能力或不同的通信延迟)时,同步和循环操作效率很低,因为它们会使工作程序空闲以等待较慢的节点完成其计算。在本文中,我们针对具有平滑损失函数的正则化随机优化问题提出了一种异步微型批处理算法,该算法消除了空闲等待时间,并允许工作人员以最大更新率运行。我们表明,通过适当选择步长值,该算法对一般凸正则化函数的阶为O(1 /T√),对于强凸正则化函数的阶为O(1 / T),其中T是迭代次数。在这两种情况下,异步对我们算法的收敛速度的影响在渐近上都是可以忽略的,并且可以预期工人数量将接近线性加速。理论结果在分布式计算基础架构上的实际实现中得到了证实。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号