首页> 美国卫生研究院文献>Sensors (Basel Switzerland) >Dual Dynamic Scheduling for Hierarchical QoS in Uplink-NOMA: A Reinforcement Learning Approach
【2h】

Dual Dynamic Scheduling for Hierarchical QoS in Uplink-NOMA: A Reinforcement Learning Approach

机译:上行链路中的分层QoS的双动态调度:强化学习方法

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

The demand for bandwidth-intensive and delay-sensitive services is surging daily with the development of 5G technology, resulting in fierce competition for scarce radio resources. Power domain Nonorthogonal Multiple Access (NOMA) technologies can dramatically improve system capacity and spectrum efficiency. Unlike existing NOMA scheduling that mainly focuses on fairness, this paper proposes a power control solution for uplink hybrid OMA and PD-NOMA in dual dynamic environments: dynamic and imperfect channel information together with the random user-specific hierarchical quality of service (QoS). This paper models the power control problem as a nonconvex stochastic, which aims to maximize system energy efficiency while guaranteeing hierarchical user QoS requirements. Then, the problem is formulated as a partially observable Markov decision process (POMDP). Owing to the difficulty of modeling time-varying scenes, the urgency of fast convergency, the adaptability in a dynamic environment, and the continuity of the variables, a Deep Reinforcement Learning (DRL)-based method is proposed. This paper also transforms the hierarchical QoS constraint under the NOMA serial interference cancellation (SIC) scene to fit DRL. The simulation results verify the effectiveness and robustness of the proposed algorithm under a dual uncertain environment. As compared with the baseline Particle Swarm Optimization algorithm (PSO), the proposed DRL-based method has demonstrated satisfying performance.
机译:随着5G技术的发展,对带宽密集型和延迟敏感服务的需求正在日常飙升,导致稀缺无线电资源激烈竞争。功率域非正交多址(NOMA)技术可以显着提高系统容量和频谱效率。与主要关注公平性的现有NOMA调度不同,本文提出了双重动态环境中的上行链路混合OMA和PD-NOMA的电源控制解决方案:动态和不完美的信道信息以及随机用户特定的分层服务(QoS)。本文将功率控制问题模拟为非透露随机性,旨在最大限度地提高系统能效,同时保证分层用户QoS要求。然后,将问题标准为部分观察到的马尔可夫决策过程(POMDP)。由于造型随时间变化的场景的难度,快速收敛的紧迫性,在动态环境中的适应性和变量的连续性,深强化学习(DRL)为基础的方法提出。本文还将NOMA串行干扰消除(SIC)场景下的分层QoS约束转换为适合DRL。仿真结果验证了在双重不确定环境下所提出的算法的有效性和鲁棒性。与基线粒子群优化算法(PSO)相比,所提出的基于DRL的方法表明了满足性能。

著录项

  • 期刊名称 Sensors (Basel Switzerland)
  • 作者单位
  • 年(卷),期 2021(21),13
  • 年度 2021
  • 页码 4404
  • 总页数 12
  • 原文格式 PDF
  • 正文语种
  • 中图分类
  • 关键词

    机译:深度确定性政策梯度(DDPG);分层QoS;非正交多通道(NOMA);功率分配;强化学习(RL);
  • 入库时间 2022-08-21 12:34:39

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号