【24h】

Thompson Sampling for Optimizing Stochastic Local Search

机译:汤普森抽样优化随机局部搜索

获取原文

摘要

Stochastic local search (SLS), like many other stochastic optimization algorithms, has several parameters that need to be optimized in order for the algorithm to find high quality solutions within a short amount of time. In this paper, we formulate a stochastic local search bandit (SLSB), which is a novel learning variant of SLS based on multi-armed bandits. SLSB optimizes SLS over a sequence of stochastic optimization problems and achieves high average cumulative reward. In SLSB, we study how SLS can be optimized via low degree polynomials in its noise and restart parameters. To determine the coefficients of the polynomials, we present polynomial Thompson Sampling (PolyTS). We derive a regret bound for PolyTS and validate its performance on synthetic problems of varying difficulty as well as on feature selection problems. Compared to bandits with no assumptions of the reward function and other parameter optimization approaches, our PolyTS assuming polynomial structure can provide substantially better parameter optimization for SLS. In addition, due to its simple model update, PolyTS has low computational cost compared to other SLS parameter optimization methods.
机译:像许多其他随机优化算法一样,随机局部搜索(SLS)具有几个需要优化的参数,以便算法在短时间内找到高质量的解决方案。在本文中,我们制定了一种随机局部搜索土匪(SLSB),它是一种基于多臂土匪的SLS的新型学习变体。 SLSB通过一系列随机优化问题对SLS进行优化,并获得较高的平均累积奖励。在SLSB中,我们研究如何通过低阶多项式在其噪声和重启参数中优化SLS。为了确定多项式的系数,我们提出多项式汤普森采样(PolyTS)。我们对PolyTS感到遗憾,并验证了其在各种难度的综合问题以及特征选择问题上的性能。与没有假设奖励函数和其他参数优化方法的土匪相比,我们的假设多项式结构的PolyTS可以为SLS提供更好的参数优化。此外,由于其简单的模型更新,与其他SLS参数优化方法相比,PolyTS的计算成本较低。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号