首页> 外文期刊>Journal of Parallel and Distributed Computing >Proactive scheduling in distributed computing-A reinforcement learning approach
【24h】

Proactive scheduling in distributed computing-A reinforcement learning approach

机译:分布式计算中的主动调度-一种强化学习方法

获取原文
获取原文并翻译 | 示例
           

摘要

In distributed computing such as grid computing, online users submit their tasks anytime and anywhere to dynamic resources. Task arrival and execution processes are stochastic. How to adapt to the consequent uncertainties, as well as scheduling overhead and response time, are the main concern in dynamic scheduling. Based on the decision theory, scheduling is formulated as a Markov decision process (MDP). To address this problem, an approach from machine learning is used to learn task arrival and execution patterns online. The proposed algorithm can automatically acquire such knowledge without any aforehand modeling, and proactively allocate tasks on account of the forthcoming tasks and their execution dynamics. Under comparison with four classic algorithms such as Min-Min, Min-Max, Suffrage, and ECT, the proposed algorithm has much less scheduling overhead. The experiments over both synthetic and practical environments reveal that the proposed algorithm outperforms other algorithms in terms of the average response time. The smaller variance of average response time further validates the robustness of our algorithm.
机译:在诸如网格计算之类的分布式计算中,在线用户可以随时随地向动态资源提交任务。任务到达和执行过程是随机的。如何适应随之而来的不确定性以及调度开销和响应时间,是动态调度中的主要问题。基于决策理论,调度被表述为马尔可夫决策过程(MDP)。为了解决这个问题,使用了机器学习的方法来在线学习任务到达和执行模式。所提出的算法可以自动获取这种知识,而无需任何预先建模,并且基于即将到来的任务及其执行动态来主动分配任务。与Min-Min,Min-Max,Suffrage和ECT等四种经典算法相比,该算法的调度开销要小得多。在综合和实际环境中进行的实验表明,在平均响应时间方面,该算法优于其他算法。平均响应时间的较小方差进一步验证了我们算法的鲁棒性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号