首页> 外文会议>Annual conference on Neural Information Processing Systems >Inverse Reinforcement Learning through Structured Classification
【24h】

Inverse Reinforcement Learning through Structured Classification

机译:通过结构化分类进行逆强化学习

获取原文

摘要

This paper adresses the inverse reinforcement learning (IRL) problem, that is inferring a reward for which a demonstrated expert behavior is optimal. We introduce a new algorithm, SCIRL, whose principle is to use the so-called feature expectation of the expert as the parameterization of the score function of a multi-class classifier. This approach produces a reward function for which the expert policy is provably near-optimal. Contrary to most of existing IRL algorithms, SCIRL does not require solving the direct RL problem. Moreover, with an appropriate heuristic, it can succeed with only trajectories sampled according to the expert behavior. This is illustrated on a car driving simulator.
机译:本文解决了逆向强化学习(IRL)问题,即推断出对已证明的专家行为最佳的奖励。我们介绍了一种新算法SCIRL,其原理是使用专家的所谓特征期望作为多类分类器得分函数的参数化。这种方法产生了一个奖励函数,专家策略证明该函数是接近最优的。与大多数现有的IRL算法相反,SCIRL不需要解决直接RL问题。此外,使用适当的启发式方法,仅根据专家行为对轨迹进行采样即可成功。汽车驾驶模拟器上对此进行了说明。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号