首页> 外文期刊>Computational Optimization and Applications >A combined class of self-scaling and modified quasi-Newton methods
【24h】

A combined class of self-scaling and modified quasi-Newton methods

机译:自定标和改进的拟牛顿法的组合类

获取原文
       

摘要

Techniques for obtaining safely positive definite Hessian approximations with self-scaling and modified quasi-Newton updates are combined to obtain ‘better’ curvature approximations in line search methods for unconstrained optimization. It is shown that this class of methods, like the BFGS method, has the global and superlinear convergence for convex functions. Numerical experiments with this class, using the well-known quasi-Newton BFGS, DFP and a modified SR1 updates, are presented to illustrate some advantages of the new techniques. These experiments show that the performance of several combined methods are substantially better than that of the standard BFGS method. Similar improvements are also obtained if the simple sufficient function reduction condition on the steplength is used instead of the strong Wolfe conditions.
机译:结合使用具有自定标和修正的拟牛顿更新的安全正定Hessian逼近技术,可以在无限制优化的直线搜索方法中获得“更好”的曲率逼近。结果表明,这类方法与BFGS方法一样,对于凸函数具有全局和超线性收敛性。此类的数值实验使用著名的准牛顿BFGS,DFP和改进的SR1更新进行了演示,以说明新技术的一些优点。这些实验表明,几种组合方法的性能明显优于标准BFGS方法。如果使用步长上的简单充分函数约简条件代替强Wolfe条件,也会获得类似的改进。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号