首页> 外文期刊>Human-Machine Systems, IEEE Transactions on >Human-Centered Reinforcement Learning: A Survey
【24h】

Human-Centered Reinforcement Learning: A Survey

机译:以人为本的强化学习:一项调查

获取原文
获取原文并翻译 | 示例
           

摘要

Human-centered reinforcement learning (RL), in which an agent learns how to perform a task from evaluative feedback delivered by a human observer, has become more and more popular in recent years. The advantage of being able to learn from human feedback for a RL agent has led to increasing applicability to real-life problems. This paper describes the state-of-the-art human-centered RL algorithms and aims to become a starting point for researchers who are initiating their endeavors in human-centered RL. Moreover, the objective of this paper is to present a comprehensive survey of the recent breakthroughs in this field and provide references to the most interesting and successful works. After starting with an introduction of the concepts of RL from environmental reward, this paper discusses the origins of human-centered RL and its difference from traditional RL. Then we describe different interpretations of human evaluative feedback, which have produced many human-centered RL algorithms in the past decade. In addition, we describe research on agents learning from both human evaluative feedback and environmental rewards as well as on improving the efficiency of human-centered RL. Finally, we conclude with an overview of application areas and a discussion of future work and open questions.
机译:近年来,以人为中心的强化学习(RL)变得越来越流行,在该学习中,代理人通过人类观察者提供的评估反馈来学习如何执行任务。能够从人为RL代理的反馈中学习的优势已导致对现实生活中问题的适用性增加。本文介绍了最先进的以人为中心的RL算法,旨在成为那些开始以人为中心的RL研究工作的研究人员的起点。此外,本文的目的是对这一领域的最新进展进行全面的综述,并为最有趣,最成功的著作提供参考。在从环境奖励中介绍RL概念之后,本文讨论了以人为中心的RL的起源及其与传统RL的区别。然后,我们描述对人类评价反馈的不同解释,这些解释在过去十年中产生了许多以人为中心的RL算法。此外,我们描述了有关从人类评估反馈和环境奖励中学习的行为主体以及提高以人为本的RL效率的研究。最后,我们以应用领域的概述以及对未来工作和未解决问题的讨论作为结尾。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号