首页> 外文会议>American Control Conference >Probabilistic density control for swarm of decentralized ON-OFF agents with safety constraints
【24h】

Probabilistic density control for swarm of decentralized ON-OFF agents with safety constraints

机译:具有安全约束的分散式ON-OFF代理群的概率密度控制

获取原文
获取外文期刊封面目录资料

摘要

This paper presents a Markov chain based approach for the probabilistic density control of a swarm of autonomous “ON-OFF” agents. The proposed approach specifies the time evolution of the probabilistic density distribution by using a Markov chain, which guides the swarm to a desired steady-state final distribution, while satisfying the prescribed ergodicity and safety constraints. Prior research has developed a Markov chain based approach to control swarms of agents with full mobility. The main contribution of the current paper is generalizing this approach to a swarm of ON-OFF agents with limited mobility. We define ON-OFF agents as having limited mobility in the following sense: The agent either conforms to the motion induced by the environment or it remains motionless. This means that an ON-OFF agent has two possible actions, either accept the environmentally induced motion, “ON”, or stop all the motion, “OFF”. By using these binary control actions at the agent level, we develop a decentralized control architecture and algorithms that guide the swarm density distribution to a desired probabilistic density in the operational space. The agents make statistically independent probabilistic decisions on choosing to be “ON” or “OFF” based solely on their own states to achieve a desired swarm density distribution. The probabilistic approach is completely decentralized and does not require communication or collaboration between agents. Of course, any collaboration can be leveraged for better performance, which is the subject of future work. There are two new algorithms developed: An online ON-OFF policy computation method to generate a Markov matrix with the ergodicity and motion constraints but without the safety constraints, which can be viewed as generating a Markov matrix via the Metropolis-Hastings (M-H) algorithm for a given proposal matrix. The second algorithm generates, offline, an - N-OFF policy that also ensures the safety constraints together with the ergodicity and motion constraints. The incorporation of the safety constraints is enabled by our recent result that convexifies the Markov chain synthesis with these constraints.
机译:本文提出了一种基于马尔可夫链的方法,用于自动“开-关”智能体群的概率密度控制。所提出的方法通过使用马尔可夫链来指定概率密度分布的时间演化,该马尔可夫链在满足规定的遍历性和安全性约束的同时,将群体引导至所需的稳态最终分布。先前的研究已经开发出一种基于马尔可夫链的方法来控制具有完全移动性的特工群体。当前论文的主要贡献是将这种方法推广到了行动不便的大量ON-OFF代理上。在以下意义上,我们将ON-OFF代理定义为具有受限的移动性:代理要么符合环境诱导的运动,要么保持静止。这意味着ON-OFF代理有两种可能的动作,要么接受环境引起的运动“ ON”,要么停止所有运动“ OFF”。通过在代理程序级别使用这些二进制控制动作,我们开发了分散的控制体系结构和算法,可将群密度分布引导到操作空间中的所需概率密度。代理程序仅根据其自身状态来选择是否为“开”或“关”即可做出统计上独立的概率决策,以实现所需的群密度分布。概率方法是完全分散的,不需要代理之间的沟通或协作。当然,可以利用任何协作来获得更好的性能,这是未来工作的主题。开发了两种新算法:一种在线开-关策略计算方法,以生成具有遍历和运动约束但没有安全约束的马尔可夫矩阵,可以将其视为通过Metropolis-Hastings(MH)算法生成马尔可夫矩阵给定的提案矩阵。第二种算法离线生成-N-OFF策略,该策略还确保安全约束以及遍历性和运动约束。我们最近的结果使安全约束的合并成为可能,这些结果使马尔可夫链综合具有这些约束。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号