...
首页> 外文期刊>Intelligent and Converged Networks >Deep reinforcement learning for dynamic computation offloading and resource allocation in cache-assisted mobile edge computing systems
【24h】

Deep reinforcement learning for dynamic computation offloading and resource allocation in cache-assisted mobile edge computing systems

机译:强化学习动态计算卸载和资源分配cache-assisted移动计算系统

获取原文
获取原文并翻译 | 示例

摘要

Mobile Edge Computing (MEC) is one of the most promising techniques for next-generation wireless communication systems. In this paper, we study the problem of dynamic caching, computation offloading, and resource allocation in cache-assisted multi-user MEC systems with stochastic task arrivals. There are multiple computationally intensive tasks in the system, and each Mobile User (MU) needs to execute a task either locally or remotely in one or more MEC servers by offloading the task data. Popular tasks can be cached in MEC servers to avoid duplicates in offloading. The cached contents can be either obtained through user offloading, fetched from a remote cloud, or fetched from another MEC server. The objective is to minimize the long-term average of a cost function, which is defined as a weighted sum of energy consumption, delay, and cache contents' fetching costs. The weighting coefficients associated with the different metrics in the objective function can be adjusted to balance the tradeoff among them. The optimum design is performed with respect to four decision parameters: whether to cache a given task, whether to offload a given uncached task, how much transmission power should be used during offloading, and how much MEC resources to be allocated for executing a task. We propose to solve the problems by developing a dynamic scheduling policy based on Deep Reinforcement Learning (DRL) with the Deep Deterministic Policy Gradient (DDPG) method. A new decentralized DDPG algorithm is developed to obtain the optimum designs for multi-cell MEC systems by leveraging on the cooperations among neighboring MEC servers. Simulation results demonstrate that the proposed algorithm outperforms other existing strategies, such as Deep Q-Network (DQN).
机译:移动边缘计算(MEC)是一种最有前途的下一代无线技术通信系统。动态缓存的问题,计算卸载和资源分配MEC cache-assisted多用户系统随机任务到达。计算密集型任务系统中,每个移动用户(μ)需要执行一项任务在一个或多个MEC在本地或远程服务器通过卸载任务数据。任务可以缓存MEC服务器来避免重复出售。是通过用户卸载,从远程获取云,或获取另一个MEC服务器。的长期平均成本函数,该函数被定义为一个加权和的能量消费、延迟和缓存内容的抓取成本。不同指标的目标函数可以调整平衡之间的权衡他们。对四个参数:决定是否缓存一个给定的任务,是否卸载不缓存任务,传动功率应该多少在卸载,MEC多少资源分配给执行一个任务。我们建议通过开发一个解决这个问题基于深度的动态调度策略强化学习(DRL)深决定性策略梯度(DDPG)方法。新的分散DDPG算法开发获得多单元MEC的最佳设计系统通过利用之间的合作邻近MEC服务器。证明该算法优于现有的其他策略,如深Q-Network (DQN)。

著录项

相似文献

  • 外文文献
  • 中文文献
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号