首页> 外文期刊>Journal of supercomputing >Availability-aware and energy-aware dynamic SFC placement using reinforcement learning
【24h】

Availability-aware and energy-aware dynamic SFC placement using reinforcement learning

机译:可用性感知和能量感知动态SFC使用强化学习的展示

获取原文
获取原文并翻译 | 示例
       

摘要

Software-defined networking and network functions virtualisation are making networks programmable and consequently much more flexible and agile. To meet service-level agreements, achieve greater utilisation of legacy networks, faster service deployment, and reduce expenditure, telecommunications operators are deploying increasingly complex service function chains (SFCs). Notwithstanding the benefits of SFCs, increasing heterogeneity and dynamism from the cloud to the edge introduces significant SFC placement challenges, not least adding or removing network functions while maintaining availability, quality of service, and minimising cost. In this paper, an availability- and energy-aware solution based on reinforcement learning (RL) is proposed for dynamic SFC placement. Two policy-aware RL algorithms, Advantage Actor-Critic (A2C) and Proximal Policy Optimisation (PPO), are compared using simulations of a ground truth network topology based on the Rede Nacional de Ensino e Pesquisa Network, Brazil's National Teaching and Research Network backbone. The simulation results show that PPO generally outperformed A2C and a greedy approach in terms of both acceptance rate and energy consumption. The biggest difference in the PPO when compared to the other algorithms relates to the SFC availability requirement of 99.965%; the PPO algorithm median acceptance rate is 67.34% better than the A2C algorithm. A2C outperforms PPO only in the scenario where network servers had a greater number of computing resources. In this case, the A2C is 1% better than the PPO.
机译:软件定义的网络和网络功能虚拟化正在制作网络可编程,从而更灵活,灵活。为了满足服务级别协议,实现遗留网络的更大利用,更快的服务部署,减少支出,电信运营商正在部署越来越复杂的服务功能链(SFC)。尽管SFC的好处,但从云到边缘的不均匀性和动态度引入了显着的SFC放置挑战,并不是保持或移除网络功能,同时保持可用性,服务质量和最小化成本。本文提出了一种基于加强学习(RL)的可用性和能量感知解决方案,用于动态SFC放置。使用基于Rede Nacional De Insinoine Pesquisa网络,巴西国家教学和研究网络骨干的地面真实网络拓扑模拟比较了两个策略感知RL算法,优势演员 - 评论家(A2C)和近端政策优化(PPO)。 。仿真结果表明,在接受率和能量消耗方面,PPO通常优于A2C和贪婪的方法。与其他算法相比,PPO的最大差异涉及99.965%的SFC可用性要求; PPO算法中值接受率比A2C算法更好67.34%。 A2C仅在网络服务器具有更多计算资源的情况下才能表达PPO。在这种情况下,A2C比PPO好1%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号