首页> 外文期刊>Artificial Intelligence Review >Research on task decomposition and state abstraction in reinforcement learning
【24h】

Research on task decomposition and state abstraction in reinforcement learning

机译:强化学习中的任务分解与状态抽象研究

获取原文
           

摘要

Task decomposition and State abstraction are crucial parts in reinforcement learning. It allows an agent to ignore aspects of its current states that are irrelevant to its current decision, and therefore speeds up dynamic programming and learning. This paper presents the SVI algorithm that uses a dynamic Bayesian network model to construct an influence graph that indicates relationships between state variables. SVI performs state abstraction for each subtask by ignoring irrelevant state variables and lower level subtasks. Experiment results show that the decomposition of tasks introduced by SVI can significantly accelerate constructing a near-optimal policy. This general framework can be applied to a broad spectrum of complex real world problems such as robotics, industrial manufacturing, games and others.
机译:任务分解和状态抽象是强化学习的关键部分。它使代理可以忽略与其当前决策无关的当前状态方面,从而加快了动态编程和学习的速度。本文介绍了使用动态贝叶斯网络模型构建表示状态变量之间关系的影响图的SVI算法。 SVI通过忽略无关的状态变量和较低级别的子任务来为每个子任务执行状态抽象。实验结果表明,SVI引入的任务分解可以显着加快构建接近最优的策略。该通用框架可应用于各种复杂的现实世界问题,例如机器人技术,工业制造,游戏等。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号