首页> 外文期刊>Journal of robotics >Dynamic Path Planning of Unknown Environment Based on Deep Reinforcement Learning
【24h】

Dynamic Path Planning of Unknown Environment Based on Deep Reinforcement Learning

机译:基于深度强化学习的未知环境动态路径规划

获取原文
           

摘要

Dynamic path planning of unknown environment has always been a challenge for mobile robots. In this paper, we apply double Q-network (DDQN) deep reinforcement learning proposed by DeepMind in 2016 to dynamic path planning of unknown environment. The reward and punishment function and the training method are designed for the instability of the training stage and the sparsity of the environment state space. In different training stages, we dynamically adjust the starting position and target position. With the updating of neural network and the increase of greedy rule probability, the local space searched by agent is expanded. Pygame module in PYTHON is used to establish dynamic environments. Considering lidar signal and local target position as the inputs, convolutional neural networks (CNNs) are used to generalize the environmental state. Q-learning algorithm enhances the ability of the dynamic obstacle avoidance and local planning of the agents in environment. The results show that, after training in different dynamic environments and testing in a new environment, the agent is able to reach the local target position successfully in unknown dynamic environment.
机译:未知环境的动态路径规划一直是移动机器人面临的挑战。在本文中,我们将DeepMind在2016年提出的双Q网络(DDQN)深度强化学习应用于未知环境的动态路径规划。针对训练阶段的不稳定和环境状态空间的稀疏性设计了奖惩功能和训练方法。在不同的训练阶段,我们会动态调整起始位置和目标位置。随着神经网络的更新和贪婪规则概率的增加,智能体搜索的局部空间得以扩展。 PYTHON中的Pygame模块用于建立动态环境。考虑到激光雷达信号和本地目标位置作为输入,使用卷积神经网络(CNN)概括环境状态。 Q学习算法增强了环境中智能体的动态避障能力和局部计划能力。结果表明,在不同的动态环境中进行训练并在新的环境中进行测试后,代理可以在未知的动态环境中成功到达本地目标位置。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号