首页> 外文会议>World Congress on Intelligent Control and Automation >An Improved Q-learning Algorithm Based on Exploration Region Expansion Strategy
【24h】

An Improved Q-learning Algorithm Based on Exploration Region Expansion Strategy

机译:一种改进的基于勘探区域扩展策略的Q学习算法

获取原文

摘要

In order to find a good solution to one of the key problems in Q-learning algorithm - keeping the balance between exploration and exploitation, an improved Q-learning algorithm based on exploration region expansion strategy is proposed on the base of Metropolis criterion-based Q-learning. With this strategy, the exploration blindness in the entire environment is eliminated, and the learning efficiency is increased. Meanwhile, other feasible path is sought where agent encounters obstacles, which makes the implementation of the algorithm on real robot easy. An automatic termination condition is also put forward, therefore, the redundant learning after finding optimal path is avoided, and the time of learning is reduced. The validity of the algorithm is proved by simulation experiments.
机译:为了找到Q-Learnal算法中的一个关键问题的良好解决方案 - 保持勘探和开发之间的平衡,提出了一种基于勘探区域扩展策略的改进的Q学习算法,基于基于大都市标准的Q基础-学习。通过这种策略,消除了整个环境中的勘探失明,并且增加了学习效率。同时,寻求代理遇到障碍物的其他可行路径,这使得实际机器人的实现变得容易。因此,还提出了自动终止条件,因此避免了找到最佳路径之后的冗余学习,并且减少了学习的时间。通过模拟实验证明了算法的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号