首页> 外文期刊>Foundations and trends in information retrieval >Bandit Algorithms in Information Retrieval
【24h】

Bandit Algorithms in Information Retrieval

机译:信息检索中的强盗算法

获取原文
       

摘要

Bandit algorithms, named after casino slot machines sometimes known as "one-armed bandits", fall into a broad category of stochastic scheduling problems. In the setting with multiple arms, each arm generates a reward with a given probability. The gambler's aim is to find the arm producing the highest payoff and then continue playing in order to accumulate the maximum reward possible. However, having only a limited number of plays, the gambler is faced with a dilemma: should he play the arm currently known to produce the highest reward or should he keep on trying other arms in the hope of finding a better paying one? This problem formulation is easily applicable to many real-life scenarios, hence in recent years there has been an increased interest in developing bandit algorithms for a range of applications. In information retrieval and recommender systems, bandit algorithms, which are simple to implement and do not require any training data, have been particularly popular in online personalization, online ranker evaluation and search engine optimization. This survey provides a brief overview of bandit algorithms designed to tackle specific issues in information retrieval and recommendation and, where applicable, it describes how they were applied in practice.
机译:以娱乐场老虎机(有时被称为“单臂土匪”)命名的强盗算法属于随机调度问题的大类。在多臂情况下,每个臂都会以给定的概率生成奖励。赌徒的目的是找到产生最高回报的手臂,然后继续玩以积累最大可能的报酬。但是,赌徒数量有限,因此面临一个难题:他应该玩目前已知能产生最高报酬的那支武器,还是应该继续尝试其他武器以希望找到薪水更高的武器?这个问题的表达很容易适用于许多现实生活中的场景,因此,近年来,对于开发适用于各种应用的强盗算法的兴趣日益浓厚。在信息检索和推荐系统中,易于实现且不需要任何训练数据的匪徒算法在在线个性化,在线排名评估和搜索引擎优化中特别受欢迎。此调查简要概述了旨在解决信息检索和推荐中的特定问题的强盗算法,并在适用时描述了它们在实践中的应用方式。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号