...
首页> 外文期刊>Artificial life and robotics >Developing reinforcement learning for adaptive co-construction of continuous high-dimensional state and action spaces
【24h】

Developing reinforcement learning for adaptive co-construction of continuous high-dimensional state and action spaces

机译:开发增强学习以适应连续高维状态和动作空间的自适应共建

获取原文
获取原文并翻译 | 示例
           

摘要

Engineers and researchers are paying more attention to reinforcement learning (RL) as a key technique for realizing adaptive and autonomous decentralized systems. In general, however, it is not easy to put RL into practical use. Our approach mainly deals with the problem of designing state and action spaces. Previously, an adaptive state space construction method which is called a "state space filter" and an adaptive action space construction method which is called "switching RL", have been proposed after the other space has been fixed. Then, we have reconstituted these two construction methods as one method by treating the former method and the latter method as a combined method for mimicking an infant's perceptual and motor developments and we have proposed a method which is based on introducing and referring to "entropy". In this paper, a computational experiment was conducted using a so-called "robot navigation problem" with three-dimensional continuous state space and two-dimensional continuous action space which is more complicated than a so-called "path planning problem". As a result, the validity of the proposed method has been confirmed.
机译:工程师和研究人员越来越重视强化学习(RL),这是实现自适应和自治分散系统的关键技术。但是,一般而言,将RL投入实际使用并不容易。我们的方法主要处理设计状态和动作空间的问题。先前,已经提出了在固定了另一空间之后的被称为“状态空间滤波器”的自适应状态空间构造方法和被称为“切换RL”的自适应动作空间构造方法。然后,我们通过将前一种方法和后一种方法作为模仿婴儿的知觉和运动发育的组合方法,将这两种构建方法重构为一种方法,并提出了一种基于引入和引用“熵”的方法。在本文中,使用具有三维连续状态空间和二维连续作用空间的所谓“机器人导航问题”进行了计算实验,该问题比所谓的“路径规划问题”更为复杂。结果,已经证实了所提出的方法的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号