...
首页> 外文期刊>Wireless Communications Letters, IEEE >Dynamic Pricing for Smart Mobile Edge Computing: A Reinforcement Learning Approach
【24h】

Dynamic Pricing for Smart Mobile Edge Computing: A Reinforcement Learning Approach

机译:智能移动边缘计算的动态定价:强化学习方法

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

This letter studies the revenue maximization problem for the mobile edge computing (MEC) system, where an access point (AP) is equipped with an MEC server, providing job offloading service for multiple resource-hungry users and charging users a service fee for it. Usually, the information about users' personal demand is unknown and users' job arrival rate is time-varying, which make pricing highly challenging. As such, we develop a policy gradient (PG)-based reinforcement learning (RL) algorithm. In specific, a deep neural network (DNN) is adopted as the policy network to design price policy, and a baseline neural network (BNN) is used to reduce the inherent high variance of the gradient obtained using PG. The proposed PG-based algorithm enables continuous pricing, thus constituting an advancement over the conventional Q-learning algorithm that has provided only discrete action space. Simulation results show that our proposed method converges to the optimal revenue performance, while the Q-learning algorithm suffers 44% revenue loss.
机译:这封信研究了移动边缘计算(MEC)系统的收入最大化问题,其中接入点(AP)配备了MEC服务器,为多个资源饥饿的用户提供工作卸载服务,并为用户充电服务费用。通常,有关用户个人需求的信息是未知的,用户的工作到达率是时变的,这使得定价高度挑战。因此,我们开发了一种策略梯度(PG)基础的加强学习(RL)算法。具体地,采用深度神经网络(DNN)作为策略网络来设计价格政策,并且基线神经网络(BNN)用于降低使用PG获得的梯度的固有高方差。所提出的基于PG的算法使得能够连续定价,从而构成了仅提供了仅提供了离散动作空间的传统Q学习算法的进步。仿真结果表明,我们所提出的方法会聚到最佳收入性能,而Q学习算法遭受44%的收入损失。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号