...
首页> 外文期刊>International Journal of Control, Automation and Systems >An efficient initialization approach of Q-learning for mobile robots
【24h】

An efficient initialization approach of Q-learning for mobile robots

机译:一种高效的移动机器人Q学习初始化方法

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

This article demonstrates that Q-learning can be accelerated by appropriately specifying initial Q-values using dynamic wave expansion neural network. In our method, the neural network has the same topography as robot work space. Each neuron corresponds to a certain discrete state. Every neuron of the network will reach an equilibrium state according to the initial environment information. The activity of the special neuron denotes the maximum cumulative reward by following the optimal policy from the corresponding state when the network is stable. Then the initial Q-values are defined as the immediate reward plus the maximum cumulative reward by following the optimal policy beginning at the succeeding state. In this way, we create a mapping between the known environment information and the initial values of Q-table based on neural network. The prior knowledge can be incorporated into the learning system, and give robots a better learning foundation. Results of experiments in a grid world problem show that neural network-based Q-learning enables a robot to acquire an optimal policy with better learning performance compared to conventional Q-learning and potential field-based Qlearning.
机译:本文证明,通过使用动态波扩展神经网络适当指定初始Q值,可以加速Q学习。在我们的方法中,神经网络具有与机器人工作空间相同的地形。每个神经元对应于某个离散状态。网络的每个神经元将根据初始环境信息达到平衡状态。当网络稳定时,通过遵循来自相应状态的最佳策略,特殊神经元的活动表示最大的累积奖励。然后,通过遵循从后续状态开始的最佳策略,将初始Q值定义为立即奖励加上最大累积奖励。通过这种方式,我们基于神经网络在已知环境信息和Q表的初始值之间创建映射。可以将先验知识整合到学习系统中,并为机器人提供更好的学习基础。网格世界问题的实验结果表明,与传统的Q学习和潜在的基于领域的Q学习相比,基于神经网络的Q学习使机器人能够获得具有更好的学习性能的最优策略。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号