首页> 外文期刊>ISA Transactions >Reinforcement learning algorithms for robotic navigation in dynamic environments
【24h】

Reinforcement learning algorithms for robotic navigation in dynamic environments

机译:动态环境中机器人导航的强化学习算法

获取原文
获取原文并翻译 | 示例
获取外文期刊封面目录资料

摘要

The purpose of this study was to examine improvements to reinforcement learning (RL) algorithms in order to successfully interact within dynamic environments. The scope of the research was that of RL algorithms as applied to robotic navigation. Proposed improvements include: addition of a forgetting mechanism, use of feature based state inputs, and hierarchical structuring of an RL agent. Simulations were performed to evaluate the individual merits and flaws of each proposal, to compare proposed methods to prior established methods, and to compare proposed methods to theoretically optimal solutions. Incorporation of a forgetting mechanism did considerably improve the learning times of RL agents in a dynamic environment. However, direct implementation of a feature-based RL agent did not result in any performance enhancements, as pure feature-based navigation results in a lack of positional awareness, and the inability of the agent to determine the location of the goal state. Inclusion of a hierarchical structure in an RL agent resulted in significantly improved performance, specifically when one layer of the hierarchy included a feature-based agent for obstacle avoidance, and a standard RL agent for global navigation. In summary, the inclusion of a forgetting mechanism, and the use of a hierarchically structured RL agent offer substantially increased performance when compared to traditional RL agents navigating in a dynamic environment.
机译:这项研究的目的是研究对强化学习(RL)算法的改进,以便在动态环境中成功进行交互。研究的范围是应用于机器人导航的RL算法。拟议的改进包括:添加遗忘机制,使用基于功能的状态输入以及RL代理的分层结构。进行了仿真,以评估每个建议的优点和缺点,将建议的方法与先前建立的方法进行比较,并将建议的方法与理论上最佳的解决方案进行比较。引入遗忘机制的确大大改善了动态环境中RL代理的学习时间。但是,直接执行基于功能的RL代理程序不会导致任何性能增强,因为基于纯功能的导航会导致缺乏位置意识,并且代理程序无法确定目标状态的位置。在RL代理中包含层次结构可显着提高性能,特别是当层次的一层包括用于避障的基于特征的代理和用于全局导航的标准RL代理时。总之,与在动态环境中导航的传统RL代理相比,包括遗忘机制以及使用分层结构的RL代理可以显着提高性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号