首页> 外文期刊>IEEE transactions on wireless communications >Energy-Efficient Mode Selection and Resource Allocation for D2D-Enabled Heterogeneous Networks: A Deep Reinforcement Learning Approach
【24h】

Energy-Efficient Mode Selection and Resource Allocation for D2D-Enabled Heterogeneous Networks: A Deep Reinforcement Learning Approach

机译:能源有效的D2D异构网络选择和资源分配:深度加强学习方法

获取原文
获取原文并翻译 | 示例
           

摘要

Improving energy efficiency has shown increasing importance in designing future cellular system. In this work, we consider the issue of energy efficiency in D2D-enabled heterogeneous cellular networks. Specifically, communication mode selection and resource allocation are jointly considered with the aim to maximize the energy efficiency in the long term. And an Markov decision process (MDP) problem is formulated, where each user can switch between traditional cellular mode and D2D mode dynamically. We employ deep deterministic policy gradient (DDPG), a model-free deep reinforcement learning algorithm, to solve the MDP problem in continuous state and action space. The architecture of proposed method consists of one actor network and one critic network. The actor network uses deterministic policy gradient scheme to generate deterministic actions for agent directly, and the critic network employs value function based Q networks to evaluate the performance of the actor network. Simulation results show the convergence property of proposed algorithm and the effectiveness in improving the energy efficiency in a D2D-enabled heterogeneous network.
机译:提高能源效率表明在设计未来蜂窝系统方面的重要性越来越重要。在这项工作中,我们考虑了能效在D2D的异构蜂窝网络中的能效问题。具体地,连续考虑通信模式选择和资源分配,其目的是长期最大化能量效率。和制定马尔可夫决策过程(MDP)问题,其中每个用户可以动态地在传统的蜂窝模式和D2D模式之间切换。我们采用深度确定性政策梯度(DDPG),一种无模型的深度增强学习算法,以解决连续状态和动作空间的MDP问题。建议方法的架构包括一个演员网络和一个批评网络。 Actor网络使用确定性策略梯度方案直接为代理生成确定性动作,并且批评网络采用基于价值函数的Q网络来评估演员网络的性能。仿真结果显示了所提出的算法的收敛性及改善D2D异构网络中能量效能的效果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号