首页> 外文期刊>Foundations and trends in information retrieval >Bandit Algorithms in Information Retrieval
【24h】

Bandit Algorithms in Information Retrieval

机译:信息检索中的强盗算法

获取原文
       

摘要

Bandit algorithms, named after casino slot machines sometimes known as "one-armed bandits", fall into a broad category of stochastic scheduling problems. In the setting with multiple arms, each arm generates a reward with a given probability. The gambler's aim is to find the arm producing the highest payoff and then continue playing in order to accumulate the maximum reward possible. However, having only a limited number of plays, the gambler is faced with a dilemma: should he play the arm currently known to produce the highest reward or should he keep on trying other arms in the hope of finding a better paying one? This problem formulation is easily applicable to many real-life scenarios, hence in recent years there has been an increased interest in developing bandit algorithms for a range of applications. In information retrieval and recommender systems, bandit algorithms, which are simple to implement and do not require any training data, have been particularly popular in online personalization, online ranker evaluation and search engine optimization. This survey provides a brief overview of bandit algorithms designed to tackle specific issues in information retrieval and recommendation and, where applicable, it describes how they were applied in practice.
机译:Bandit算法,以有时称为“单武器匪徒”的赌场老虎机命名,属于广泛的随机调度问题。在具有多个臂的设置中,每个臂产生具有给定概率的奖励。赌徒的目标是找到生产最高收益的手臂,然后继续播放,以累积最大奖励。但是,只有有限数量的戏剧,赌徒面临困境:他是否应该扮演目前已知的手臂来产生最高奖励,或者他应该继续尝试其他武器,希望找到一个更好的支付一个人?该问题配方很容易适用于许多现实生活场景,因此近年来在开发了一系列应用中开发强盗算法的兴趣增加。在信息检索和推荐系统中,强盗算法,易于实施并且不需要任何培训数据,在线个性化,在线排名评估和搜索引擎优化方面特别受欢迎。本调查介绍了匪盗算法概述,旨在解决信息检索和推荐中的特定问题,并且在适用的情况下,它描述了它们在实践中的应用方式。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号