首页> 外文OA文献 >A Taylor polynomial expansion line search for large-scale optimization
【2h】

A Taylor polynomial expansion line search for large-scale optimization

机译:泰勒多项式展开线搜索用于大规模优化

摘要

In trying to cope with the Big Data deluge, the landscape of distributed computing has changed. Large commodity hardware clusters, typically operating in some form of MapReduce framework, are becoming prevalent for organizations that require both tremendous storage capacity and fault tolerance. However, the high cost of communication can dominate the computation time in large-scale optimization routines in these frameworks. This thesis considers the problem of how to efficiently conduct univariate line searches in commodity clusters in the context of gradient-based batch optimization algorithms, like the staple limited-memory BFGS (LBFGS) method. In it, a new line search technique is proposed for cases where the underlying objective function is analytic, as in logistic regression and low rank matrix factorization. The technique approximates the objective function by a truncated Taylor polynomial along a fixed search direction. The coefficients of this polynomial may be computed efficiently in parallel with far less communication than needed to transmit the high-dimensional gradient vector, after which the polynomial may be minimized with high accuracy in a neighbourhood of the expansion point without distributed operations. This Polynomial Expansion Line Search (PELS) may be invoked iteratively until the expansion point and minimum are sufficiently accurate, and can provide substantial savings in time and communication costs when multiple iterations in the line search procedure are required.Three applications of the PELS technique are presented herein for important classes of analytic functions: (i) logistic regression (LR), (ii) low-rank matrix factorization (MF) models, and (iii) the feedforward multilayer perceptron (MLP). In addition, for LR and MF, implementations of PELS in the Apache Spark framework for fault-tolerant cluster computing are provided. These implementations conferred significant convergence enhancements to their respective algorithms, and will be of interest to Spark and Hadoop practitioners. For instance, the Spark PELS technique reduced the number of iterations and time required by LBFGS to reach terminal training accuracies for LR models by factors of 1.8--2. Substantial acceleration was also observed for the Nonlinear Conjugate Gradient algorithm for MLP models, which is an interesting case for future study in optimization for neural networks. The PELS technique is applicable to a broad class of models for Big Data processing and large-scale optimization, and can be a useful component of batch optimization routines.
机译:为了应对大数据洪水,分布式计算的格局已经改变。大型商品硬件集群通常以某种形式的MapReduce框架运行,对于同时需要巨大存储容量和容错能力的组织来说,这种集群正变得越来越普遍。但是,在这些框架中,高昂的通信成本可能会占据大规模优化例程中的计算时间。本文考虑了如何在基于梯度的批量优化算法(如装订有限内存BFGS(LBFGS)方法)的背景下有效地进行商品集群中的单变量线搜索的问题。其中,针对逻辑目标回归和低秩矩阵分解的情况,针对潜在目标函数被解析的情况,提出了一种新的线搜索技术。该技术通过沿固定搜索方向的截断泰勒多项式近似目标函数。可以以比传输高维梯度向量所需的通信少得多的通信并行地有效地并行计算该多项式的系数,此后可以在扩展点附近以高精度将多项式最小化而无需分布式操作。可以迭代调用此多项式扩展线搜索(PELS),直到扩展点和最小值足够准确为止,并且当需要在线搜索过程中进行多次迭代时,可以大大节省时间和通信成本。PELS技术的三个应用是本文针对重要的分析功能类别提供了以下内容:(i)Logistic回归(LR),(ii)低秩矩阵分解(MF)模型和(iii)前馈多层感知器(MLP)。此外,对于LR和MF,还提供了Apache Spark框架中用于容错群集计算的PELS实现。这些实现为它们各自的算法带来了显着的融合增强,Spark和Hadoop从业者将对此感兴趣。例如,Spark PELS技术将LBFGS达到LR模型的最终训练精度所需的迭代次数和时间减少了1.8--2倍。对于MLP模型的非线性共轭梯度算法也观察到了显着的加速,这是未来神经网络优化研究的一个有趣案例。 PELS技术适用于大数据处理和大规模优化的各种模型,并且可以作为批处理优化例程的有用组成部分。

著录项

  • 作者

    Hynes Michael`;

  • 作者单位
  • 年度 2016
  • 总页数
  • 原文格式 PDF
  • 正文语种 en
  • 中图分类

相似文献

  • 外文文献
  • 中文文献
  • 专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号