首页> 外文期刊>Internet of Things Journal, IEEE >Reinforcement Learning for Decentralized Trajectory Design in Cellular UAV Networks With Sense-and-Send Protocol
【24h】

Reinforcement Learning for Decentralized Trajectory Design in Cellular UAV Networks With Sense-and-Send Protocol

机译:带有感知和发送协议的蜂窝无人机网络中分散轨迹设计的强化学习

获取原文
获取原文并翻译 | 示例
           

摘要

Recently, the unmanned aerial vehicles (UAVs) have been widely used in real-time sensing applications over cellular networks. The performance of a UAV is determined by both its sensing and transmission processes, which are influenced by the trajectory of the UAV. However, it is challenging for the UAV to determine its trajectory, since it works in a dynamic environment, where other UAVs determine their trajectories dynamically and compete for the limited spectrum resources in the same time. To tackle this challenge, we adopt the reinforcement learning to solve the UAV trajectory design problem in a decentralized manner. To coordinate multiple UAVs performing real-time sensing tasks, we first propose a sense-and-send protocol, and analyze the probability for successful valid data transmission using nested Markov chains. Then, we propose an enhanced multi-UAV Q-learning algorithm to solve the decentralized UAV trajectory design problem. Simulation results show that the proposed algorithm converges faster and achieves higher utilities for the UAVs, compared to traditional single-and multi-agent Q-learning algorithms.
机译:最近,无人飞行器(UAV)已被广泛用于蜂窝网络的实时感测应用中。无人机的性能取决于其感测和传输过程,受无人机飞行轨迹的影响。但是,由于无人机要在动态环境中工作,因此无人机要确定其轨迹具有挑战性,在该环境中其他无人机可以动态确定其轨迹并同时竞争有限的频谱资源。为了应对这一挑战,我们采用强化学习的方式来分散解决无人机航迹设计问题。为了协调执行实时传感任务的多架无人机,我们首先提出一种传感发送协议,并使用嵌套的马尔可夫链分析成功进行有效数据传输的可能性。然后,我们提出了一种增强的多UAV Q学习算法,以解决分散的无人机航迹设计问题。仿真结果表明,与传统的单智能体和多智能体Q学习算法相比,该算法收敛更快,对无人机具有更高的实用性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号