首页> 外文期刊>Swarm and Evolutionary Computation >Synergism of Firefly Algorithm and Q-Learning for Robot Arm Path Planning
【24h】

Synergism of Firefly Algorithm and Q-Learning for Robot Arm Path Planning

机译:萤火虫算法的协同作用与机器人臂路径规划的Q学习

获取原文
获取原文并翻译 | 示例
       

摘要

Over the past few decades, Firefly Algorithm (FA) has attracted the attention of many researchers by virtue of its capability of solving complex real-world optimization problems. The only factor restricting the efficiency of this FA algorithm is the need of having balanced exploration and exploitation while searching for the global optima in the search-space. This balance can be established by tuning the two inherent control parameters of FA. One is the randomization parameter and another is light absorption coefficient, over iterations, either experimentally or by an automatic adaptive strategy. This paper aims at the later by proposing an improvised FA which involves the Q-learning framework within itself. In this proposed Q-learning induced FA (QFA), the optimal parameter values for each firefly of a population are learnt by the Q-learning strategy during the learning phase and applied thereafter during execution. The proposed algorithm has been simulated on fifteen benchmark functions suggested in the CEC 2015 competition. In addition, the proposed algorithm's superiority is tested by conducting the Friedman test, Iman Davenport and Bonferroni Dunn test. Moreover, its suitability for application in real-world constrained environments has been examined by employing the algorithm in the path planning of a robotic manipulator amidst various obstacles. To avoid obstacles one mechanism is designed for the robot-arm. The results, obtained from both simulation and real-world experiment, confirm the superiority of the proposed QFA over other contender algorithms in terms of solution quality as well as run-time complexity.
机译:在过去的几十年里,萤火虫算法(FA)凭借解决复杂的真实优化问题,吸引了许多研究人员的注意。限制该FA算法效率的唯一因素是需要具有平衡的探索和剥削,同时搜索搜索空间中的全局最优。可以通过调整FA的两个固有控制参数来建立这种平衡。一个是随机化参数,另一个是在实验或通过自动自适应策略的迭代的光吸收系数。本文通过提出即兴的FA提出涉及本身内的Q学习框架的简历FA。在该提出的Q学习诱导的FA(QFA)中,在学习阶段期间Q学习策略学习群体的每个萤火虫的最佳参数值,并在执行期间应用。在CEC 2015竞争中建议的十五个基准函数模拟了所提出的算法。此外,通过进行弗里德曼测试,IMAN DAVENPORT和BONFERRONI DUNN测试来测试所提出的算法的优越性。此外,它通过在各种障碍物中使用机器人操纵器的路径规划中的算法来检查其在实际约束环境中应用的适用性。为避免障碍,为机器人手臂设计了一种机制。从模拟和现实世界的实验中获得的结果,在解决方案质量以及运行时复杂性方面证实了所提出的QFA在其他竞争者算法上的优越性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号