首页> 外文会议>IEEE International symposium on Personal, Indoor, and Mobile Radio Communications >Q-learning based power control algorithm for D2D communication
【24h】

Q-learning based power control algorithm for D2D communication

机译:基于Q学习的D2D通信功率控制算法

获取原文

摘要

In this paper, reinforcement learning (RL) based power control algorithm in underlay D2D communication is studied. The approach we use regards D2D communication as a multi-agents system, and power control is achieved by maximizing system capacity while maintaining the requirement of quality of service(QoS) from cellular users. We propose two RL based power control methods for D2D users, i.e., team-Q learning and distributed-Q learning. The former is a centralized method in which only one Q-value table needs to be maintained, while the latter enables D2D users to learn independently and reduces the complexity of Q-value table. Simulation results show the difference of the two Q-learning algorithm in terms of convergence and reward function. In addition, it is shown that through our distributed-Q learning, D2D users not only are able to learn their power in a self-organized way, but also achieve better system performance than that using traditional method in LTE(Long Term Evolution).
机译:本文研究了底层D2D通信中基于强化学习(RL)的功率控制算法。我们使用的方法将D2D通信视为多代理系统,并且在保持蜂窝用户对服务质量(QoS)的要求的同时,通过最大化系统容量来实现功率控制。我们为D2D用户提出了两种基于RL的功率控制方法,即Team-Q学习和Distributed-Q学习。前者是一种集中式方法,其中只需要维护一个Q值表,而后者则使D2D用户能够独立学习并降低Q值表的复杂性。仿真结果表明两种Q学习算法在收敛性和奖励函数方面的差异。此外,还表明,通过我们的分布式Q学习,D2D用户不仅能够以自组织的方式学习他们的能力,而且还可以获得比LTE(长期演进)中使用传统方法更好的系统性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号