首页> 外文期刊>Future generation computer systems >Energy-aware resource management for uplink non-orthogonal multiple access: Multi-agent deep reinforcement learning
【24h】

Energy-aware resource management for uplink non-orthogonal multiple access: Multi-agent deep reinforcement learning

机译:用于上行链路非正交多路访问的能源感知资源管理:多主体深度强化学习

获取原文
获取原文并翻译 | 示例
           

摘要

Non-orthogonal multiple access (NOMA) is one of the promising technologies to meet the huge access demand and the high data rate requirements of the next generation networks. In this paper, we investigate the joint subchannel assignment and power allocation problem in an uplink multi-user NOMA system to maximize the energy efficiency (EE) while ensuring the quality-of-service (QoS) of all users. Different from conventional model-based resource allocation methods, we propose two deep reinforcement learning (DRL) based frameworks to solve this non-convex and dynamic optimization problem, referred to as discrete DRL based resource allocation (DDRA) framework and continuous DRL based resource allocation (CDRA) framework. Specifically, for the DDRA framework, we use a deep Q network (DQN) to output the optimum subchannel assignment policy, and design a distributed and discretized multi-DQN based network to allocate the corresponding transmit power of all users. For the CDRA framework, we design a joint DQN and deep deterministic policy gradient (DDPG) based network to generate the optimal subchannel assignment and power allocation policy. The entire resource allocation policies of these two frameworks are adjusted by updating the weights of their neural networks according to feedback of the system. Numerical results show that the proposed DRL-based resource allocation frameworks can significantly improve the EE of the whole NOMA system compared with other approaches. The proposed DRL based frameworks can provide good performance in various moving speed scenarios through adjusting learning parameters.
机译:非正交多路访问(NOMA)是满足下一代网络巨大的访问需求和高数据速率要求的有前途的技术之一。在本文中,我们研究了上行链路多用户NOMA系统中的联合子信道分配和功率分配问题,以在确保所有用户的服务质量(QoS)的同时最大化能效(EE)。与传统的基于模型的资源分配方法不同,我们提出了两种基于深度强化学习(DRL)的框架来解决此非凸和动态优化问题,分别称为基于离散DRL的资源分配(DDRA)框架和基于连续DRL的资源分配(CDRA)框架。具体来说,对于DDRA框架,我们使用深度Q网络(DQN)输出最佳子信道分配策略,并设计一个基于分布式和离散化的基于多DQN的网络,以分配所有用户的相应发射功率。对于CDRA框架,我们设计了一个基于DQN和深度确定性策略梯度(DDPG)的联合网络,以生成最佳子信道分配和功率分配策略。通过根据系统的反馈更新其神经网络的权重,可以调整这两个框架的整个资源分配策略。数值结果表明,与其他方法相比,基于DRL的资源分配框架可以显着提高整个NOMA系统的EE。所提出的基于DRL的框架可以通过调整学习参数在各种移动速度情况下提供良好的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号