首页> 外文会议>International Conference on Learning and Intelligent Optimization >Rover Descent: Learning to Optimize by Learning to Navigate on Prototypical Loss Surfaces
【24h】

Rover Descent: Learning to Optimize by Learning to Navigate on Prototypical Loss Surfaces

机译:流浪者下降:通过学习在原型损失面上进行导航来学习优化

获取原文

摘要

Learning to optimize - the idea that we can learn from data algorithms that optimize a numerical criterion - has recently been at the heart of a growing number of research efforts. One of the most challenging issues within this approach is to learn a policy that is able to optimize over classes of functions that are different from the classes that the policy was trained on. We propose a novel way of framing learning to optimize as a problem of learning a good navigation policy on a partially observable loss surface. To this end, we develop Rover Descent, a solution that allows us to learn a broad optimization policy from training only on a small set of prototypical two-dimensional surfaces that encompasses classically hard cases such as valleys, plateaus, cliffs and saddles and by using strictly zeroth-order information. We show that, without having access to gradient or curvature information, we achieve fast convergence on optimization problems not presented at training time, such as the Rosenbrock function and other two-dimensional hard functions. We extend our framework to optimize over high dimensional functions and show good preliminary results.
机译:学习优化-我们可以从优化数值准则的数据算法中学习的思想-最近成为越来越多的研究工作的核心。这种方法中最具挑战性的问题之一是学习一种策略,该策略能够针对与该策略所针对的功能类别不同的​​功能类别进行优化。我们提出了一种框架学习的新方法,以优化作为在部分可观察到的损失表面上学习良好导航策略的问题。为此,我们开发了Rover Descent,该解决方案使我们能够通过仅在少量原型二维表面上进行训练来学习广泛的优化策略,其中包括经典的困难情况,例如山谷,高原,悬崖和马鞍,并通过使用严格为零阶信息。我们表明,无需访问梯度或曲率信息,就可以在训练时未出现的优化问题上实现快速收敛,例如Rosenbrock函数和其他二维硬函数。我们扩展了框架,以优化高维函数并显示出良好的初步结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号