首页> 外文期刊>JMLR: Workshop and Conference Proceedings >Learning to Learn without Gradient Descent by Gradient Descent
【24h】

Learning to Learn without Gradient Descent by Gradient Descent

机译:通过梯度下降学习无梯度下降的学习

获取原文
           

摘要

We learn recurrent neural network optimizers trained on simple synthetic functions by gradient descent. We show that these learned optimizers exhibit a remarkable degree of transfer in that they can be used to efficiently optimize a broad range of derivative-free black-box functions, including Gaussian process bandits, simple control objectives, global optimization benchmarks and hyper-parameter tuning tasks. Up to the training horizon, the learned optimizers learn to trade-off exploration and exploitation, and compare favourably with heavily engineered Bayesian optimization packages for hyper-parameter tuning.
机译:我们学习递归神经网络优化器,通过梯度下降对简单的合成函数进行训练。我们表明,这些博学的优化器表现出非凡的转移程度,因为它们可用于有效地优化广泛的无导数黑盒函数,包括高斯过程强盗,简单的控制目标,全局优化基准和超参数调整任务。在培训之前,有学识的优化人员将学会权衡探索和开发,并与用于超级参数调整的精心设计的贝叶斯优化软件包进行比较。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号