首页> 外文会议>International conference on Machine learning >Experience-efficient learning in associative bandit problems
【24h】

Experience-efficient learning in associative bandit problems

机译:联想土匪问题中的经验有效学习

获取原文

摘要

We formalize the associative bandit problem framework introduced by Kaelbling as a learning-theory problem. The learning environment is modeled as a k-armed bandit where arm payoffs are conditioned on an observable input selected on each trial. We show that, if the payoff functions are constrained to a known hypothesis class, learning can be performed efficiently with respect to the VC dimension of this class. We formally reduce the problem of PAC classification to the associative bandit problem, producing an efficient algorithm for any hypothesis class for which efficient classification algorithms are known. We demonstrate the approach empirically on a scalable concept class.
机译:我们将由Kaelbling引入的联想土匪问题框架正式化为学习理论问题。学习环境被建模为 k 武装匪徒,其中武装收益取决于每次试验中选择的可观察输入。我们证明,如果将支付函数限制在已知的假设类中,则可以相对于该类的VC维有效地执行学习。我们正式将PAC分类问题简化为相关的匪徒问题,从而为任何已知有效分类算法的假设类别提供了一种有效的算法。我们在可扩展的概念类上以经验方式演示了该方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号