首页> 美国卫生研究院文献>Sensors (Basel Switzerland) >On-Demand Channel Bonding in Heterogeneous WLANs: A Multi-Agent Deep Reinforcement Learning Approach
【2h】

On-Demand Channel Bonding in Heterogeneous WLANs: A Multi-Agent Deep Reinforcement Learning Approach

机译:异构WLAN中的按需信道绑定:多代理深度强化学习方法

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

In order to meet the ever-increasing traffic demand of Wireless Local Area Networks (WLANs), channel bonding is introduced in IEEE 802.11 standards. Although channel bonding effectively increases the transmission rate, the wider channel reduces the number of non-overlapping channels and is more susceptible to interference. Meanwhile, the traffic load differs from one access point (AP) to another and changes significantly depending on the time of day. Therefore, the primary channel and channel bonding bandwidth should be carefully selected to meet traffic demand and guarantee the performance gain. In this paper, we proposed an On-Demand Channel Bonding (O-DCB) algorithm based on Deep Reinforcement Learning (DRL) for heterogeneous WLANs to reduce transmission delay, where the APs have different channel bonding capabilities. In this problem, the state space is continuous and the action space is discrete. However, the size of action space increases exponentially with the number of APs by using single-agent DRL, which severely affects the learning rate. To accelerate learning, Multi-Agent Deep Deterministic Policy Gradient (MADDPG) is used to train O-DCB. Real traffic traces collected from a campus WLAN are used to train and test O-DCB. Simulation results reveal that the proposed algorithm has good convergence and lower delay than other algorithms.
机译:为了满足无线局域网(WLAN)不断增长的流量需求,在IEEE 802.11标准中引入了信道绑定。尽管信道绑定有效地提高了传输速率,但是较宽的信道会减少非重叠信道的数量,并且更容易受到干扰。同时,流量负载从一个接入点(AP)到另一个接入点(AP)有所不同,并且会根据一天中的时间发生显着变化。因此,应仔细选择主要信道和信道绑定带宽,以满足流量需求并保证性能增益。在本文中,我们针对异构WLAN提出了一种基于深度强化学习(DRL)的按需信道绑定(O-DCB)算法,以减少传输延迟,其中AP具有不同的信道绑定功能。在这个问题中,状态空间是连续的,而动作空间是离散的。但是,通过使用单代理DRL,动作空间的大小会随着AP数量的增加而呈指数增长,这严重影响了学习速度。为了加速学习,多代理深度确定性策略梯度(MADDPG)用于训练O-DCB。从校园WLAN收集的真实流量跟踪用于训练和测试O-DCB。仿真结果表明,与其他算法相比,该算法具有较好的收敛性和较低的时延。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号