首页> 外文期刊>JMLR: Workshop and Conference Proceedings >Variational inference for the multi-armed contextual bandit
【24h】

Variational inference for the multi-armed contextual bandit

机译:多臂上下文强盗的变分推理

获取原文
           

摘要

In many biomedical, science, and engineering problems, one must sequentially decide which action to take next so as to maximize rewards. One general class of algorithms for optimizing interactions with the world, while simultaneously learning how the world operates, is the multi-armed bandit setting and, in particular, the contextual bandit case. In this setting, for each executed action, one observes rewards that are dependent on a given ’context’, available at each interaction with the world. The Thompson sampling algorithm has recently been shown to enjoy provable optimality properties for this set of problems, and to perform well in real-world settings. It facilitates generative and interpretable modeling of the problem at hand. Nevertheless, the design and complexity of the model limit its application, since one must both sample from the distributions modeled and calculate their expected rewards. We here show how these limitations can be overcome using variational inference to approximate complex models, applying to the reinforcement learning case advances developed for the inference case in the machine learning community over the past two decades. We consider contextual multi-armed bandit applications where the true reward distribution is unknown and complex, which we approximate with a mixture model whose parameters are inferred via variational inference. We show how the proposed variational Thompson sampling approach is accurate in approximating the true distribution, and attains reduced regrets even with complex reward distributions. The proposed algorithm is valuable for practical scenarios where restrictive modeling assumptions are undesirable.
机译:在许多生物医学,科学和工程学问题中,必须依次决定下一步应该采取的行动,以使报酬最大化。一类用于优化与世界互动的算法,同时学习世界的运行方式的多类算法是多臂匪徒设置,尤其是上下文匪徒情况。在这种设置下,对于每个执行的动作,都会观察到与给定“上下文”相关的奖励,这种奖励可在与世界的每次互动中获得。汤普森采样算法最近被证明具有针对这组问题的可证明的最优性,并且在实际环境中表现良好。它有助于对当前问题进行生成和可解释的建模。但是,模型的设计和复杂性限制了它的应用,因为必须从建模的分布中采样并计算其预期收益。我们在这里展示了如何使用变分推理来近似复杂模型来克服这些限制,并将其应用于过去二十年来针对机器学习社区中的推理案例开发的强化学习案例。我们考虑真实奖励分配未知且复杂的上下文多臂强盗应用程序,我们使用混合模型进行近似,该模型的参数是通过变分推断来推断的。我们展示了拟议的变分汤普森采样方法在逼近真实分布时如何准确,即使复杂的奖励分布也能减少后悔。所提出的算法对于不希望使用限制性建模假设的实际情况非常有价值。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号