...
首页> 外文期刊>IFAC PapersOnLine >Deep Reinforcement Learning and Randomized Blending for Control under Novel Disturbances
【24h】

Deep Reinforcement Learning and Randomized Blending for Control under Novel Disturbances

机译:新型扰动下控制的深度加固学习和随机混合

获取原文

摘要

Enabling autonomous vehicles to maneuver in novel scenarios is a key unsolved problem. A well-known approach, Weighted Multiple Model Adaptive Control (WMMAC), uses a set of pre-tuned controllers and combines their control actions using a weight vector. Although WMMAC offers an improvement to traditional switched control in terms of smooth control oscillations, it depends on accurate fault isolation and cannot deal with unknown disturbances. A recent approach avoids state estimation by randomly assigning the controller weighting vector; however, this approach uses a uniform distribution for control-weight sampling, which is sub-optimal compared to state-estimation methods. In this article, we propose a framework that uses deep reinforcement learning (DRL) to learn weighted control distributions that optimize the performance of the randomized approach for both known and unknown disturbances. We show that RL-based randomized blending dominates pure randomized blending, a switched FDI-based architecture and pre-tuned controllers on a quadcopter trajectory optimisation task in which we penalise deviations in both position and attitude.
机译:在新颖的情景中启用自动车辆的机动是一个关键未解决的问题。众所周知的方法,加权多模型自适应控制(WMMAC)使用一组预调谐控制器,并使用权重向量组合它们的控制动作。虽然WMMAC在平滑控制振荡方面对传统的交换控制提供了改进,但它取决于精确的故障隔离,不能处理未知的干扰。最近的方法通过随机分配控制器加权矢量来避免状态估计;然而,与状态估计方法相比,这种方法使用用于控制权重采样的均匀分布,这是次优的。在本文中,我们提出了一个使用深度加强学习(DRL)来学习加权控制分布的框架,从而优化随机方法的性能,以了解已知和未知干扰。我们表明基于RL的随机混合在Quadcopter轨迹优化任务上占据了纯随机混合,纯粹的基于FDI的架构和预调谐控制器,在其中惩罚了两个位置和姿态的偏差。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号