首页> 外文期刊>Automatica Sinica, IEEE/CAA Journal of >Feature-based aggregation and deep reinforcement learning: a survey and some new implementations
【24h】

Feature-based aggregation and deep reinforcement learning: a survey and some new implementations

机译:基于特征的聚合和深度强化学习:一项调查和一些新的实现

获取原文
获取原文并翻译 | 示例
           

摘要

In this paper we discuss policy iteration methods for approximate solution of a finite-state discounted Markov decision problem, with a focus on feature-based aggregation methods and their connection with deep reinforcement learning schemes. We introduce features of the states of the original problem, and we formulate a smaller “aggregate” Markov decision problem, whose states relate to the features. We discuss properties and possible implementations of this type of aggregation, including a new approach to approximate policy iteration. In this approach the policy improvement operation combines feature-based aggregation with feature construction using deep neural networks or other calculations. We argue that the cost function of a policy may be approximated much more accurately by the nonlinear function of the features provided by aggregation, than by the linear function of the features provided by neural network-based reinforcement learning, thereby potentially leading to more effective policy improvement.
机译:在本文中,我们讨论了用于有限状态折扣Markov决策问题的近似解的策略迭代方法,重点是基于特征的聚合方法及其与深度强化学习方案的联系。我们介绍了原始问题的状态特征,并制定了一个较小的“汇总”马尔可夫决策问题,其状态与特征有关。我们讨论了这种聚合的属性和可能的​​实现,包括一种近似策略迭代的新方法。在这种方法中,策略改进操作将基于特征的聚合与使用深度神经网络或其他计算的特征构建相结合。我们认为,与通过基于神经网络的强化学习提供的特征的线性函数相比,通过聚集提供的特征的非线性函数可以更准确地近似策略的成本函数,从而有可能导致更有效的策略改善。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号