...
首页> 外文期刊>IEEE Transactions on Vehicular Technology >A New Block-Based Reinforcement Learning Approach for Distributed Resource Allocation in Clustered IoT Networks
【24h】

A New Block-Based Reinforcement Learning Approach for Distributed Resource Allocation in Clustered IoT Networks

机译:集群式物联网网络中基于资源的基于块的强化学习新方法

获取原文
获取原文并翻译 | 示例
           

摘要

Resource allocation and spectrum management are two major challenges in the massive scale deployment of Internet of Things (IoT) and Machine-to-Machine (M2M) communication. Furthermore, the large number of devices per unit area in IoT networks also leads to congestion, network overload, and deterioration of the Signal to Noise Ratio (SNR). To address these problems, efficient resource allocation play a pivotal role in optimizing the throughput, delay, and power management of IoT networks. To this end, most of the existing resource allocation mechanisms are centralized and do not gracefully support the heterogeneous and dynamic IoT networks. Therefore, distributed and Machine Learning (ML)-based approaches are essential. However, distributed resource allocation techniques also have scalability problem with large number of devices whereas the ML-based approaches are currently scarce in the literature. In this paper, we propose a new distributed block-based Q-learning algorithm for slot scheduling in the smart devices and Machine Type Communication Devices (MTCDs) participating in clustered IoT networks. We furthermore, propose various reward schemes for the evolution of Q-values in the proposed scheme and, discuss and evaluate their effect on the distributed model. Our goal is to avoid inter- and intra-cluster interference, and to improve the Signal to Interference Ratio (SIR) by employing frequency diversity in a multi-channel system. Through extensive simulations, we analyze the effects of the distributed slot-assignment (with respect to varying SIR) on the convergence rate and the convergence probability. Our theoretical analysis and simulations validate the effectiveness of our proposed method where, (i) a suitable slot with acceptable SIR levels is allocated to each MTCD, and (ii) IoT network can efficiently converge to a collision-free transmission causing minimum intra-cluster interference. The network convergence is achieved through each MTCD's learning ability during the distributed slot allocation.
机译:资源分配和频谱管理是物联网(IoT)和机器对机器(M2M)通信的大规模部署中的两个主要挑战。此外,物联网网络中每单位面积的大量设备还导致拥塞,网络过载以及信噪比(SNR)下降。为了解决这些问题,有效的资源分配在优化IoT网络的吞吐量,延迟和电源管理方面起着关键作用。为此,大多数现有资源分配机制都是集中式的,不能很好地支持异构和动态物联网网络。因此,基于分布式和机器学习(ML)的方法至关重要。然而,分布式资源分配技术还具有大量设备的可扩展性问题,而基于ML的方法目前在文献中很少。在本文中,我们提出了一种新的基于块的分布式Q学习算法,用于参与集群式IoT网络的智能设备和机器类型通信设备(MTCD)中的时隙调度。此外,我们为提出的方案中的Q值演化提出了各种奖励方案,并讨论和评估了它们对分布式模型的影响。我们的目标是避免集群之间和集群内部的干扰,并通过在多信道系统中采用频率分集来提高信噪比(SIR)。通过广泛的仿真,我们分析了分布式时隙分配(相对于变化的SIR)对收敛速度和收敛概率的影响。我们的理论分析和仿真验证了我们提出的方法的有效性,其中(i)为每个MTCD分配了一个具有可接受SIR级别的合适插槽,并且(ii)IoT网络可以有效地收敛到无冲突传输,从而使集群内部最少干扰。通过在分布式时隙分配期间每个MTCD的学习能力来实现网络融合。

著录项

  • 来源
    《IEEE Transactions on Vehicular Technology》 |2020年第3期|2891-2904|共14页
  • 作者

  • 作者单位

    Royal Bank Canada Toronto ON M5J 0B8 Canada;

    Innopolis Univ Inst Informat Secur & Cyber Phys Syst Innopolis 420500 Russia;

    Ryerson Univ Dept Comp Sci WINCORE Lab Toronto ON M5B 2K3 Canada;

    Univ Avigon F-84029 Avignon France;

  • 收录信息 美国《科学引文索引》(SCI);美国《工程索引》(EI);
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    MTCDs; clustered IoT network; machine learning; block Q-learning; resource allocation;

    机译:MTCD;集群式物联网网络机器学习阻止Q学习;资源分配;

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号