کد مقاله | کد نشریه | سال انتشار | مقاله انگلیسی | نسخه تمام متن |
---|---|---|---|---|
432726 | 689048 | 2014 | 11 صفحه PDF | دانلود رایگان |
• Propose the concept and method of proactive scheduling.
• Formulate dynamic scheduling as a MDP problem.
• Develop an online scheduling algorithm based on reinforcement learning.
• Demonstrate our learning-based algorithm stable with lower average response time.
In distributed computing such as grid computing, online users submit their tasks anytime and anywhere to dynamic resources. Task arrival and execution processes are stochastic. How to adapt to the consequent uncertainties, as well as scheduling overhead and response time, are the main concern in dynamic scheduling. Based on the decision theory, scheduling is formulated as a Markov decision process (MDP). To address this problem, an approach from machine learning is used to learn task arrival and execution patterns online. The proposed algorithm can automatically acquire such knowledge without any aforehand modeling, and proactively allocate tasks on account of the forthcoming tasks and their execution dynamics. Under comparison with four classic algorithms such as Min–Min, Min–Max, Suffrage, and ECT, the proposed algorithm has much less scheduling overhead. The experiments over both synthetic and practical environments reveal that the proposed algorithm outperforms other algorithms in terms of the average response time. The smaller variance of average response time further validates the robustness of our algorithm.
Journal: Journal of Parallel and Distributed Computing - Volume 74, Issue 7, July 2014, Pages 2662–2672