首页> 外文期刊>Machine Learning >Preference-based reinforcement learning: a formal framework and a policy iteration algorithm
【24h】

Preference-based reinforcement learning: a formal framework and a policy iteration algorithm

机译:基于偏好的强化学习:形式框架和策略迭代算法

获取原文
获取原文并翻译 | 示例
       

摘要

This paper makes a first step toward the integration of two subfields of machine learning, namely preference learning and reinforcement learning (RL). An important motivation for a preference-based approach to reinforcement learning is the observation that in many real-world domains, numerical feedback signals are not readily available, or are defined arbitrarily in order to satisfy the needs of conventional RL algorithms. Instead, we propose an alternative framework for reinforcement learning, in which qualitative reward signals can be directly used by the learner. The framework may be viewed as a generalization of the conventional RL framework in which only a partial order between policies is required instead of the total order induced by their respective expected long-term reward. Therefore, building on novel methods for preference learning, our general goal is to equip the RL agent with qualitative policy models, such as ranking functions that allow for sorting its available actions from most to least promising, as well as algorithms for learning such models from qualitative feedback. As a proof of concept, we realize a first simple instantiation of this framework that defines preferences based on utilities observed for trajectories. To that end, we build on an existing method for approximate policy iteration based on rollouts. While this approach is based on the use of classification methods for generalization and policy learning, we make use of a specific type of preference learning method called label ranking. Advantages of preference-based approximate policy iteration are illustrated by means of two case studies.
机译:本文朝着机器学习两个子领域(即偏好学习和强化学习(RL))的集成迈出了第一步。基于偏好的强化学习方法的一个重要动机是观察到,在许多实际领域中,数字反馈信号不容易获得,或者为了满足常规RL算法的需要而任意定义。相反,我们提出了一种用于强化学习的替代框架,在该框架中,学习者可以直接使用定性奖励信号。该框架可以看作是常规RL框架的概括,其中仅需要策略之间的部分顺序,而不是由它们各自的预期长期奖励引起的总顺序。因此,我们建立在偏好学习的新方法的基础上,我们的总体目标是为RL代理配备定性策略模型,例如排名功能,以将其可用操作从最大到最小排序,以及从中学习此类模型的算法。定性反馈。作为概念证明,我们实现了此框架的第一个简单实例,该实例基于对轨迹观察到的效用来定义首选项。为此,我们基于现有的方法,基于部署对策略进行近似迭代。尽管此方法基于分类方法的泛化和策略学习,但我们使用一种称为标签排名的特殊类型的偏好学习方法。通过两个案例研究说明了基于偏好的近似策略迭代的优势。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号