首页> 外文OA文献 >Machine Learning for Intelligent Control: Application of Reinforcement Learning Techniques to the Development of Flight Control Systems for Miniature UAV Rotorcraft
【2h】

Machine Learning for Intelligent Control: Application of Reinforcement Learning Techniques to the Development of Flight Control Systems for Miniature UAV Rotorcraft

机译:用于智能控制的机器学习:强化学习技术在微型无人机旋翼机飞行控制系统开发中的应用

摘要

This thesis investigates the possibility of using reinforcement learning (RL) techniques to create a flight controller for a quadrotor Micro Aerial Vehicle (MAV).A capable flight control system is a core requirement of any unmanned aerial vehicle. The challenging and diverse applications in which MAVs are destined to be used, mean that considerable time and effort need to be put into designing and commissioning suitable flight controllers. It is proposed that reinforcement learning, a subset of machine learning, could be used to address some of the practical difficulties.While much research has delved into RL in unmanned aerial vehicle applications, this work has tended to ignore low level motion control, or been concerned only in off-line learning regimes. This thesis addresses an area in which accessible information is scarce: the performance of RLwhen used for on-policy motion control.Trying out a candidate algorithm on a real MAV is a simple but expensive proposition. In place of such an approach, this research details the development of a suitable simulator environment, in which a prototype controller might be evaluated. Then inquiry then proposes a possible RL-based control system, utilising the Q-learning algorithm, with an adaptive RBF-network providing function approximation.The operation of this prototypical control system is then tested in detail, to determine both the absolute level of performance which can be expected, and the effect which tuning critical parameters of the algorithm has on the functioning of the controller. Performance is compared against a conventional PID controller to maximise the usability of the results by a wide audience. Testing considers behaviour in the presence of disturbances, and run-time changes in plant dynamics.Results show that given sufficient learning opportunity, a RL-based control system performs as well as a simple PID controller. However, unstable behaviour during learning is an issue for future analysis.Additionally, preliminary testing is performed to evaluate the feasibility of implementing RL algorithms in an embedded computing environment, as a general requirement for a MAV flight controller. Whilst the algorithm runs successfully in an embedded context, observation revealsfurther development would be necessary to reduce computation time to a level where a controller was able to update sufficiently quickly for a real-time motion control application.In summary, the study provides a critical assessment of the feasibility of using RL algorithms for motion control tasks, such as MAV flight control. Advantages which merit interest are exposed, though practical considerations suggest at this stage, that such a control system is not a realistic proposition. There is a discussion of avenues which may uncover possibilities to surmount these challenges. This investigation will prove useful for engineers interested in the opportunities which reinforcement learning techniques represent.
机译:本文研究了使用强化学习(RL)技术为四旋翼微型飞行器(MAV)创建飞行控制器的可能性。有能力的飞行控制系统是任何无人机的核心要求。注定要使用MAV的挑战性和多样化应用意味着需要在设计和调试合适的飞行控制器上投入大量的时间和精力。有人建议将强化学习作为机器学习的一个子集来解决一些实际困难。尽管在无人飞行器应用中对RL进行了大量研究,但这项工作往往忽略了低水平运动控制,或者仅关注离线学习机制。本文针对的领域是可访问信息稀缺:RL用于策略运动控制时的性能。在真实的MAV上尝试一种候选算法是一个简单但昂贵的提议。代替这种方法,本研究详细介绍了合适的模拟器环境的开发,在该环境中可以评估原型控制器。然后询问提出了一个可能的基于RL的控制系统,利用Q学习算法,通过自适应RBF网络提供函数逼近,然后详细测试该原型控制系统的运行情况,以确定绝对性能水平可以预期的结果,以及调整算法的关键参数对控制器功能的影响。将性能与常规PID控制器进行比较,以最大程度地提高广大观众对结果的可用性。测试考虑了存在干扰时的行为以及工厂动态的运行时变化。结果表明,在有足够的学习机会的情况下,基于RL的控制系统的性能与简单的PID控制器相同。然而,学习过程中的不稳定行为是未来分析的一个问题。此外,作为MAV飞行控制器的一般要求,还进行了初步测试以评估在嵌入式计算环境中实施RL算法的可行性。虽然该算法在嵌入式环境中成功运行,但观察表明,需要进一步开发,才能将计算时间减少到控制器能够为实时运动控制应用进行足够快的更新的水平。总之,本研究提供了关键的评估将RL算法用于运动控制任务(例如MAV飞行控制)的可行性。尽管在现阶段提出了实际考虑,但值得一提的优点却暴露了,这种控制系统不是现实的建议。对途径的讨论可能揭示了克服这些挑战的可能性。对于对强化学习技术所代表的机会感兴趣的工程师来说,这项研究将非常有用。

著录项

  • 作者

    Hayes Edwin Laurie;

  • 作者单位
  • 年度 2013
  • 总页数
  • 原文格式 PDF
  • 正文语种 en
  • 中图分类
  • 入库时间 2022-08-20 20:17:15

相似文献

  • 外文文献
  • 中文文献
  • 专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号