Please wait a minute...
Frontiers of Computer Science

ISSN 2095-2228

ISSN 2095-2236(Online)

CN 10-1014/TP

邮发代号 80-970

2019 Impact Factor: 1.275

Frontiers of Computer Science  2021, Vol. 15 Issue (5): 155105   https://doi.org/10.1007/s11704-020-9273-z
  本期目录
A multi-objective reinforcement learning algorithm for deadline constrained scientific workflow scheduling in clouds
Yao QIN1,2, Hua WANG3(), Shanwen YI1, Xiaole LI4(), Linbo ZHAI5
1. 1School of Computer Science and Technology, Shandong University, Jinan 250101, China
2. 2Shanghai Police College, Shanghai 200137, China
3. 3School of Software, Shandong University, Jinan 250101, China
4. 4School of Information Science and Engineering, Linyi University, Linyi 276005, China
5. 5School of Information Science and Engineering, Shandong Normal University, Jinan 250014, China
 全文: PDF(551 KB)  
Abstract

Recently, a growing number of scientific applications have been migrated into the cloud. To deal with the problems brought by clouds, more and more researchers start to consider multiple optimization goals in workflow scheduling. However, the previous works ignore some details, which are challenging but essential. Most existing multi-objective workflow scheduling algorithms overlook weight selection, which may result in the quality degradation of solutions. Besides, we find that the famous partial critical path (PCP) strategy, which has been widely used to meet the deadline constraint, can not accurately reflect the situation of each time step. Workflow scheduling is an NP-hard problem, so self-optimizing algorithms are more suitable to solve it.

In this paper, the aim is to solve a workflow scheduling problem with a deadline constraint. We design a deadline constrained scientific workflow scheduling algorithm based on multi-objective reinforcement learning (RL) called DCMORL. DCMORL uses the Chebyshev scalarization function to scalarize its Q-values. This method is good at choosing weights for objectives. We propose an improved version of the PCP strategy calledMPCP. The sub-deadlines in MPCP regularly update during the scheduling phase, so they can accurately reflect the situation of each time step. The optimization objectives in this paper include minimizing the execution cost and energy consumption within a given deadline. Finally, we use four scientific workflows to compare DCMORL and several representative scheduling algorithms. The results indicate that DCMORL outperforms the above algorithms. As far as we know, it is the first time to apply RL to a deadline constrained workflow scheduling problem.

Key wordsworkflow scheduling    energy saving    multiobjective reinforcement learning    deadline constrained    cloud computing
收稿日期: 2019-07-25      出版日期: 2021-06-11
Corresponding Author(s): Hua WANG,Xiaole LI   
 引用本文:   
. [J]. Frontiers of Computer Science, 2021, 15(5): 155105.
Yao QIN, Hua WANG, Shanwen YI, Xiaole LI, Linbo ZHAI. A multi-objective reinforcement learning algorithm for deadline constrained scientific workflow scheduling in clouds. Front. Comput. Sci., 2021, 15(5): 155105.
 链接本文:  
https://academic.hep.com.cn/fcs/CN/10.1007/s11704-020-9273-z
https://academic.hep.com.cn/fcs/CN/Y2021/V15/I5/155105
1 P K Senyo, E Addae, R Boateng. Cloud computing research: a review of research themes, frameworks, methods and future research directions. International Journal of Information Management, 2018, 38(1): 128–139
https://doi.org/10.1016/j.ijinfomgt.2017.07.007
2 A S G Andrae, T Edler. On global electricity usage of communication technology: trends to 2030. Challenges, 2015, 6(1): 117–157
https://doi.org/10.3390/challe6010117
3 J Hamilton. Cooperative expendable micro-slice servers (cems): low cost, low power servers for internet-scale services. In: Proceedings of Conference on Innovative Data Systems Research. 2009
4 N Khattar, J Sidhu, J Singh. Toward energy-efficient cloud computing: a survey of dynamic power management and heuristics-based optimization techniques. The Journal of Supercomputing, 2019, 75(8): 4750–4810
https://doi.org/10.1007/s11227-019-02764-2
5 Y Wang, H Liu, W Zheng, Y Xia, Y Li, P Chen, K Guo, H Xie. Multiobjective workflow scheduling with deep-Q-network-based multi-agent reinforcement learning. IEEE Access, 2019, 7: 39974–39982
https://doi.org/10.1109/ACCESS.2019.2902846
6 I Das, J E Dennis. A closer look at drawbacks of minimizing weighted sums of objectives for pareto set generation in multicriteria optimization problems. Structural Optimization, 1997, 14(1): 63–69
https://doi.org/10.1007/BF01197559
7 K Van Moffaert, M M Drugan, A Nowé. Scalarized multi-objective reinforcement learning: novel design techniques. In: Proceedings of IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning. 2013, 191–199
https://doi.org/10.1109/ADPRL.2013.6615007
8 S Abrishami, M Naghibzadeh, D H Epema. Deadline-constrained workflow scheduling algorithms for infrastructure as a service clouds. Future Generation Computer Systems, 2013, 29(1): 158–169
https://doi.org/10.1016/j.future.2012.05.004
9 Y Qin, H Wang, S Yi, X Li, L Zhai. An energy-aware scheduling algorithm for budget-constrained scientific workflows based on multiobjective reinforcement learning. The Journal of Supercomputing, 2020, 76: 455–480
https://doi.org/10.1007/s11227-019-03033-y
10 E Zitzler, L Thiele, M Laumanns, C M Fonseca, V G Da Fonseca. Performance assessment of multiobjective optimizers: an analysis and review. TIK-Report, 2002
https://doi.org/10.1109/TEVC.2003.810758
11 Y Qin, H Wang, S Yi, X Li, L Zhai. Virtual machine placement based on multi-objective reinforcement learning. Applied Intelligence, 2020
https://doi.org/10.1007/s10489-020-01633-3
12 R S Sutton, A G Barto. Reinforcement Learning: An Introduction. MIT Press, 2018
13 C J C H Watkins. Learning from delayed rewards. Doctoral Thesis, University of Cambridge, 1989
14 J N Tsitsiklis. Asynchronous stochastic approximation and Q-learning. Machine Learning, 1994, 16(3): 185–202
https://doi.org/10.1007/BF00993306
15 M A Wiering, E D De Jong. Computing optimal stationary policies for multi-objective markov decision processes. In: Proceedings of IEEE International Symposium on Approximate Dynamic Programming and Reinforcement Learning. 2007, 158–165
https://doi.org/10.1109/ADPRL.2007.368183
16 K Van Moffaert, A Nowé. Multi-objective reinforcement learning using sets of pareto dominating policies. The Journal of Machine Learning Research, 2014, 15(1): 3483–3512
17 P Vamplew, J Yearwood, R Dazeley, A Berry. On the limitations of scalarisation for multi-objective reinforcement learning of pareto fronts. In: Proceedings of Australasian Joint Conference on Artificial Intelligence. 2008, 372–378
https://doi.org/10.1007/978-3-540-89378-3_37
18 T Voβ, N Beume, G Rudolph, C Igel. Scalarization versus indicator-based selection in multi-objective cma evolution strategies. In: Proceedings of IEEE Congress on Evolutionary Computation (IEEE World Congress on Computational Intelligence). 2008, 3036–3043
https://doi.org/10.1109/CEC.2008.4631208
19 J Yu, R Buyya, C K Tham. Cost-based scheduling of scientific workflow applications on utility grids. In: Proceedings of the 1st International Conference on e-Science and Grid Computing. 2005
20 S Abrishami, M Naghibzadeh, D H J Epema. Cost-driven scheduling of grid workflows using partial critical paths. IEEE Transactions on Parallel and Distributed Systems, 2012, 23(8): 1400–1414
https://doi.org/10.1109/TPDS.2011.303
21 Z Li, J Ge, H Hu, W Song, H Hu, B Luo. Cost and energy aware scheduling algorithm for scientific workflows with deadline constraint in clouds. IEEE Transactions on Services Computing, 2015, 11(4): 713–726
https://doi.org/10.1109/TSC.2015.2466545
22 A Verma, S Kaushal. A hybrid multi-objective particle swarm optimization for scientific workflow scheduling. Parallel Computing, 2017, 62: 1–19
https://doi.org/10.1016/j.parco.2017.01.002
23 C A C Coello, G T Pulido, M S Lechuga. Handling multiple objectives with particle swarm optimization. IEEE Transactions on Evolutionary Computation, 2004, 8(3): 256–279
https://doi.org/10.1109/TEVC.2004.826067
24 R A Haidri, C P Katti, P C Saxena. Cost-effective deadline-aware stochastic scheduling strategy for workflow applications on virtual machines in cloud computing. Concurrency and Computation: Practice and Experience, 2019, 31(7): e5006
https://doi.org/10.1002/cpe.5006
25 K Deb, A Pratap, S Agarwal, T Meyarivan. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Transactions on Evolutionary Computation, 2002, 6(2): 182–197
https://doi.org/10.1109/4235.996017
26 J Wang, A Taal, P Martin, Y Hu, H Zhou, J Pang, d C Laat, Z Zhao. Planning virtual infrastructures for time critical applications with multiple deadline constraints. Future Generation Computer Systems, 2017, 75: 365–375
https://doi.org/10.1016/j.future.2017.02.001
27 D Zhu, R Melhem, B R Childers. Scheduling with dynamic voltage/speed adjustment using slack reclamation in multiprocessor real-time systems. IEEE Transactions on Parallel and Distributed Systems, 2003, 14(7): 686–700
https://doi.org/10.1109/TPDS.2003.1214320
28 Y C Lee, A Y Zomaya. Energy conscious scheduling for distributed computing systems under different operating conditions. IEEE Transactions on Parallel and Distributed Systems, 2010, 22(8): 1374–1381
https://doi.org/10.1109/TPDS.2010.208
29 M Atkinson, S Gesing, J Montagnat, I Taylor. Scientific workflows: past, present and future. Future Generation Computer Systems, 2017, 75: 216–227
https://doi.org/10.1016/j.future.2017.05.041
30 H Topcuoglu, S Hariri, M Y Wu. Performance-effective and lowcomplexity task scheduling for heterogeneous computing. IEEE Transactions on Parallel and Distributed Systems, 2002, 13(3): 260–274
https://doi.org/10.1109/71.993206
31 S Bharathi, A Chervenak, E Deelman, G Mehta, M H Su, K Vahi. Characterization of scientific workflows. In: Proceedings of the 3rd Workshop on Workflows in Support of Large-scale Science. 2008, 1–10
https://doi.org/10.1109/WORKS.2008.4723958
32 R N Calheiros, R Ranjan, A Beloglazov, C A De Rose, R Buyya. Cloudsim: a toolkit for modeling and simulation of cloud computing environments and evaluation of resource provisioning algorithms. Software: Practice and Experience, 2011, 41(1): 23–50
https://doi.org/10.1002/spe.995
33 N Herbst, A Bauer, S Kounev, G Oikonomou, E V Eyk, G Kousiouris, A Evangelinou, R Krebs, T Brecht, C L Abad, et al.Quantifying cloud performance and dependability: taxonomy, metric design, and emerging challenges. ACM Transactions on Modeling and Performance Evaluation of Computing Systems (ToMPECS), 2018, 3(4): 19
https://doi.org/10.1145/3236332
34 F S Melo. Convergence of Q-learning: a simple proof. Institute of Systems and Robotics, Technical Report, 2001, 1–4
[1] Article highlights Download
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed