首页> 外文会议>IEEE International Conference on Machine Learning and Applications >Multidisciplinary Optimization in Decentralized Reinforcement Learning
【24h】

Multidisciplinary Optimization in Decentralized Reinforcement Learning

机译:分散式强化学习中的多学科优化

获取原文

摘要

Multidisciplinary Optimization (MDO) is one of the most popular techniques in aerospace engineering, where the system is complex and includes the knowledge from multiple fields. However, according to the best of our knowledge, MDO has not been widely applied in decentralized reinforcement learning (RL) due to the `unknown' nature of the RL problems. In this work, we apply the MDO in decentralized RL. In our MDO design, each learning agent uses system identification to closely approximate the environment and tackle the `unknown' nature of the RL. Then, the agents apply the MDO principles to compute the control solution using Monte Carlo and Markov Decision Process techniques. We examined two options of MDO designs: the multidisciplinary feasible and the individual discipline feasible options, which are suitable for multi-agent learning. Our results show that the MDO individual discipline feasible option could successfully learn how to control the system. The MDO approach shows better performance than the completely decentralization and centralization approaches.
机译:多学科优化(MDO)是航空航天工程中最流行的技术之一,该系统非常复杂,并且包含来自多个领域的知识。但是,据我们所知,由于RL问题的“未知”性质,MDO尚未广泛应用于分散式强化学习(RL)。在这项工作中,我们将MDO应用于分散式RL。在我们的MDO设计中,每个学习代理都使用系统标识来接近环境并解决RL的“未知”性质。然后,代理将MDO原理应用到使用Monte Carlo和Markov决策过程技术来计算控制解决方案。我们研究了MDO设计的两个选项:多学科可行和个人学科可行选项,它们适用于多主体学习。我们的结果表明,MDO个人学科可行方案可以成功地学习如何控制系统。与完全分散和集中化方法相比,MDO方法显示出更好的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号