首页> 外文期刊>Concurrency and computation: practice and experience >An incremental reinforcement learning scheduling strategy for data-intensive scientific workflows in the cloud
【24h】

An incremental reinforcement learning scheduling strategy for data-intensive scientific workflows in the cloud

机译:云中数据密集型科学工作流程的增量强化学习调度策略

获取原文
获取原文并翻译 | 示例
           

摘要

Most scientific experiments can be modeled as workflows. These workflows are usually computing- and data-intensive, demanding the use of high-performance computing environments such as clusters, grids, and clouds. This latter offers the advantage of the elasticity, which allows for changing the number of virtual machines (VMs) on demand. Workflows are typically managed using scientific workflow management systems (SWfMS). Many existing SWfMSs offer support for cloud-based execution. Each SWfMS has its scheduler that follows a well-defined cost function. However, such cost functions should consider the characteristics of a dynamic environment, such as live migrations or performance fluctuations, which are far from trivial to model. This article proposes a novel scheduling strategy, named ReASSIgN, based on reinforcement learning (RL). By relying on an RL technique, one may assume that there is an optimal (or suboptimal) solution for the scheduling problem, and aims at learning the best scheduling based on previous executions in the absence of a mathematical model of the environment. For this, an extension of a well-known workflow simulator WorkflowSim is proposed to implement an RL strategy for scheduling workflows. Once the scheduling plan is generated via simulation, the workflow is executed in the cloud using SciCumulus SWfMS. We conducted a throughout evaluation of the proposed scheduling strategy using a real astronomy workflow named Montage.
机译:大多数科学实验可以被建模为工作流程。这些工作流程通常是计算和数据密集型,要求使用高性能计算环境,例如集群,网格和云。该后者提供弹性的优点,这允许根据需要改变虚拟机(VMS)的数量。工作流程通常使用科学工作流管理系统(SWFM)进行管理。许多现有的SWFMS为基于云的执行提供支持。每个SWFM都有其调度程序,遵循明确定义的成本函数。然而,这种成本函数应考虑动态环境的特征,例如实时迁移或性能波动,远远不到模型。本文提出了一种基于强化学习(RL)的名为Realsign的小说调度策略。通过依赖RL技术,可以假设对调度问题存在最佳(或次优)解决方案,并且旨在基于在不存在环境的数学模型的情况下基于先前的执行基于先前的执行来学习最佳调度。为此,提出了一个众所周知的工作流模拟器Workflowsim的扩展,以实现用于调度工作流程的RL策略。一旦通过仿真生成调度计划,就使用SciCumulus SWFMS在云中执行工作流程。我们在整个评估中,使用名为Montage的真正天文工作流程的建议的调度策略。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号