首页> 外文学位 >Improving Bayesian Optimization for Machine Learning Using Expert Priors
【24h】

Improving Bayesian Optimization for Machine Learning Using Expert Priors

机译:使用专家先验改善机器学习的贝叶斯优化

获取原文
获取原文并翻译 | 示例

摘要

Deep neural networks have recently become astonishingly successful at many machine learning problems such as object recognition and speech recognition, and they are now also being used in many new and creative ways. However, their performance critically relies on the proper setting of numerous hyperparameters. Manual tuning by an expert researcher has been a traditionally effective approach, however it is becoming increasingly infeasible as models become more complex and machine learning systems become further embedded within larger automated systems. Bayesian optimization has recently been proposed as a strategy for intelligently optimizing the hyperparameters of deep neural networks and other machine learning systems; it has been shown in many cases to outperform experts, and provides a promising way to reduce both the computational and human time required. Regardless, expert researchers can still be quite effective at hyperparameter tuning due to their ability to incorporate contextual knowledge and intuition into their search, while traditional Bayesian optimization treats each problem as a black box and therefore cannot take advantage of this knowledge. In this thesis, we draw inspiration from these abilities and incorporate them into the Bayesian optimization framework as additional prior information. These extensions include the ability to transfer knowledge between problems, the ability to transform the problem domain into one that is easier to optimize, and the ability to terminate experiments when they are no longer deemed to be promising, without requiring their training to converge. We demonstrate in experiments across a range of machine learning models that these extensions significantly reduce the cost and increase the robustness of Bayesian optimization for automatic hyperparameter tuning.
机译:深度神经网络最近在许多机器学习问题(例如对象识别和语音识别)上取得了惊人的成功,并且它们现在也以许多新颖的方式被使用。但是,它们的性能主要取决于众多超参数的正确设置。专家研究人员进行手动调整一直是传统上有效的方法,但是随着模型变得更加复杂以及机器学习系统进一步嵌入到更大的自动化系统中,这种方法变得越来越不可行。最近提出了贝叶斯优化作为智能优化深度神经网络和其他机器学习系统的超参数的策略。在许多情况下,它的性能都优于专家,并提供了一种有希望的方式来减少所需的计算时间和人工时间。无论如何,专家研究人员由于能够将上下文知识和直觉纳入其搜索中,因此仍然能够在超参数调整方面非常有效,而传统的贝叶斯优化将每个问题都视为黑匣子,因此无法利用这些知识。在本文中,我们从这些能力中汲取了灵感,并将其作为其他先验信息纳入贝叶斯优化框架中。这些扩展包括在问题之间传递知识的能力,将问题域转换为更易于优化的能力的能力,以及在不再被认为有希望的情况下终止实验的能力,而无需培训来收敛。我们在一系列机器学习模型的实验中证明,这些扩展显着降低了成本,并提高了用于自动超参数调整的贝叶斯优化的鲁棒性。

著录项

  • 作者

    Swersky, Kevin.;

  • 作者单位

    University of Toronto (Canada).;

  • 授予单位 University of Toronto (Canada).;
  • 学科 Artificial intelligence.;Computer science.
  • 学位 Ph.D.
  • 年度 2017
  • 页码 119 p.
  • 总页数 119
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号