【24h】

A novel jamming strategy-greedy bandit

机译:一种新型的干扰策略-贪婪的强盗

获取原文

摘要

In an electronic warfare-type scenario, an optimal jamming strategy is vital important for a jammer who has restricted power and how to make the optimal strategies quickly and accurately put on the agenda. In this paper, we developed a cognitive jammer who could learn the optimal jamming strategies with the proposed algorithm-Greedy Bandits (GB). By interacting with transmitter-receiver pairs continually, which is also the advantage of reinforcement learning theory, the jammer obtains the optimal physical layer parameters like signaling scheme, power level and the on-off/pulsing. After constructing the jamming model, we first prove that the proposed Greedy Bandits algorithm satisfied the jamming needs, then two new reward standard-changes in power and enduring time are also presented. Numerous results show that GB convergences more quickly than other reinforcement learning algorithm such as Jamming Bandits (JB). More importantly, GB with two proposed reward standards has an acceptable learning performance and a wide utilizing field than learning with symbol error rate (SER), despite that more interaction times is needed.
机译:在电子战类型的情况下,最佳干扰策略对于功率有限的干扰者以及如何快速,准确地将最佳策略提上议事日程至关重要。在本文中,我们开发了一种认知干扰器,可以使用提出的算法-贪婪的强盗(GB)来学习最佳干扰策略。通过与发射器-接收器对进行持续交互,这也是强化学习理论的优势,干扰器获得了最佳的物理层参数,例如信令方案,功率水平和开/关/脉冲。在构造干扰模型之后,我们首先证明了所提出的贪婪强盗算法能够满足干扰需求,然后还提出了两个新的奖励标准,即功率和持续时间的变化。许多结果表明,GB比其他强化学习算法(如Jamming Bandits(JB))的收敛速度更快。更重要的是,尽管需要更多的交互时间,但具有两种提议的奖励标准的GB与具有符号错误率(SER)的学习相比,具有可接受的学习性能和广泛的应用领域。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号