首页> 外文期刊>Internet of Things Journal, IEEE >Hierarchical Game-Theoretic and Reinforcement Learning Framework for Computational Offloading in UAV-Enabled Mobile Edge Computing Networks With Multiple Service Providers
【24h】

Hierarchical Game-Theoretic and Reinforcement Learning Framework for Computational Offloading in UAV-Enabled Mobile Edge Computing Networks With Multiple Service Providers

机译:具有多个服务提供商的支持无人机的移动边缘计算网络中用于计算分流的分层博弈论和强化学习框架

获取原文
获取原文并翻译 | 示例
       

摘要

We present a novel game-theoretic (GT) and reinforcement learning (RL) framework for computational offloading in the mobile edge computing (MEC) network operated by multiple service providers (SPs). The network is formed by MEC servers installed at stationary base stations (BSs) and unmanned aerial vehicles (UAVs) deployed as quasi-stationary BSs. Since computing powers of MEC servers are limited, the BSs in proximity can form coalitions with shared data processing resources to serve their users more efficiently. However, as BSs can be privately owned or controlled by different SPs, in any coalition, the BSs: 1) take only the actions that maximize their long-term payoffs and 2) do not coordinate their actions with other BSs in the coalition. That is, inside each coalition, BSs act in an independent and self-interested manner. Therefore, the interactions among BSs cannot be described by conventional coalitional games. Instead, the network operation is modeled by a two-level hierarchical model. The upper level is a cooperative game that defines the process of coalition formation. The lower level comprises the set of noncooperative subgames to represent a self-interested and independent behavior of BSs in coalitions. To enable each BS to select a coalition and decide on its action maximizing its long-term payoff, we propose two algorithms that combine coalition formation with RL and prove that these algorithms converge to the states where the coalitional structure is strongly stable and the strategies of BSs are in the mixed-strategy Nash equilibrium (NE).
机译:我们提出了一种新颖的博弈论(GT)和强化学习(RL)框架,用于由多个服务提供商(SP)运营的移动边缘计算(MEC)网络中的计算分流。该网络由安装在固定基站(BS)的MEC服务器和作为准固定BS部署的无人机(UAV)组成。由于MEC服务器的计算能力有限,因此邻近的BS可以与共享数据处理资源组成联盟,以更有效地为其用户提供服务。但是,由于BS可以由不同的SP私有或控制,因此在任何联盟中,BS:1)仅采取最大化其长期收益的行动; 2)不与联盟中的其他BS协调其行动。也就是说,在每个联盟内部,BS都以独立和自私的方式行动。因此,传统联盟游戏无法描述BS之间的交互。取而代之的是,网络操作由两级分层模型建模。上层是定义联盟形成过程的合作游戏。较低的级别包括代表联盟中BS的自利和独立行为的一组非合作子博弈。为了使每个BS能够选择一个联盟并决定其行动以最大化其长期收益,我们提出了两种将联盟形成与RL相结合的算法,并证明了这些算法收敛到联盟结构非常稳定的状态,并且提出了相应的策略。 BS处于混合策略纳什均衡(NE)中。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号