首页> 外文期刊>Engineering Applications of Artificial Intelligence >Adaptive state space partitioning for reinforcement learning
【24h】

Adaptive state space partitioning for reinforcement learning

机译:自适应状态空间划分,用于强化学习

获取原文
获取原文并翻译 | 示例
           

摘要

The convergence property of reinforcement learning has been extensively investigated in the field of machine learning, however, its applications to real-world problems are still constrained due to its computational complexity. A novel algorithm to improve the applicability and efficacy of reinforcement learning algorithms via adaptive state space partitioning is presented. The proposed temporal difference learning with adaptive vector quantization (TD-AVQ) is an online algorithm and does not assume any a priori knowledge with respect to the learning task and environment. It utilizes the information generated from the reinforcement learning algorithms. Therefore, no additional computations on the decisions of how to partition a particular state space are required. A series of simulations are provided to demonstrate the practical values and performance of the proposed algorithms in solving robot motion planning problems.
机译:强化学习的收敛性已经在机器学习领域中进行了广泛的研究,但是,由于其计算复杂性,它在实际问题中的应用仍然受到限制。提出了一种通过自适应状态空间划分提高强化学习算法的适用性和有效性的新算法。所提出的带有自适应矢量量化的时差学习(TD-AVQ)是一种在线算法,并且不假设有关学习任务和环境的任何先验知识。它利用了强化学习算法生成的信息。因此,不需要关于如何划分特定状态空间的决定的附加计算。提供了一系列仿真,以证明所提出算法在解决机器人运动计划问题中的实用价值和性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号