Abstract
Scientific workflows can be represented as directed acyclic graphs (DAGs) with nodes corresponding to individual tasks and directed edges between the nodes signifying the order of task execution. The nodes contain informative attributes related to task-specific data transfer/storage and scheduling length. Given an available amount of (cloud) computational resources, the overall workflow scheduling length can sometimes be reduced by making certain “critical tasks” run in parallel (task duplication) on multiple resources. In this way, a carefully designed spread of computational effort across the resources can result in a more efficient computational structure and hence a shorter scheduling length. However, task duplication algorithms deciding which tasks to duplicate in a given workflow can themselves be computationally expensive. Here we propose a novel Bottom-Up Top-Down Recursive Neural Network (BUTD RecNN) model that is able to learn from historical duplication decisions on workflows (represented as DAGs) to efficiently produce duplication recommendations for new unseen workflows. The approach is tested on collections of Montage workflows.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
The number of recursive connections of a recursive neuron should be equal to the \(max\_degree\) of the domain D, even if not all of them will be used for computing the output of a vertex v with \(out\_deg(v)< max\_degree\).
References
Berriman, G.B., et al.: Montage: a grid enabled image mosaic service for the national virtual observatory. Astronom. Data Analysis Softw. Systs. 314, 593 (2004)
Bottou, L.: From machine learning to machine reasoning. Technical report. arXiv. 1102.1808. pp. 394–396 (2011). https://doi.org/10.48550/arXiv.1102.1808
Bozdag, D., Ozguner, F., Ekici, E., Catalyurek, U.A.: task duplication based scheduling algorithm using partial schedules, In: International Conference on Parallel Processing, pp. 630–637 (2005). https://doi.org/10.1109/ICPP.2005.15
Calheiros, R. N., Ranjan, R., Beloglazov, A., De Rose, C. A. F., Buyya, R., CloudSim: a toolkit for modeling and simulation of cloud computing environments and evaluation of resource provisioning algorithms. Softw. Pract. Exper. Vol. 41, 1, pp. 23–50. (2011). https://doi.org/10.1002/spe.995
Darbha, S., Agrawal, D.P.: A fast and scalable scheduling algorithm for distributed memory systems. In: Proceedings Seventh IEEE Symposium on Parallel and Distributed Processing, pp. 60–63 (1995). https://doi.org/10.1109/SPDP.1995.530665
Deelman, E.: The future of scientific workflows. Int. J. High Performance Comput. Appl. 32(1), 159–175 (2018)
Ferreira da Silva, R., Filgueira, R., Pietri, I., Jiang, M., Sakellariou, R., Deelman, E.: A Characterization of Work-flow Management Systems for Extreme-Scale Applications. Future Gen. Comput. Syst. 5, 228–238 (2017)
Frasconi, P., Gori, M., Sperduti, A.: On the efficient classification of data structures by neural networks. In: Proceedings of the Fifteenth International Joint Conference on Artifical Intelligence - Volume 2, Morgan Kaufmann Publishers Inc., San Francisco, CA, USA, pp. 1066–1071 (1997)
Ishfaq, A., Yu-Kwong, K.: On exploiting task duplication in parallel program schedulingIEEE . Trans. Parallel Distrib. Syst. 9(9), 872–892 (1998). https://doi.org/10.1109/71.722221
Juve, G., Chervenak, A., Deelman, E., Bharathi, S., Mehta, G., Vahi, K.: Characterizing and profiling scientific workflows. Future Gen. Comput. Syst. 29(3), 682–692 (2013). https://doi.org/10.1016/j.future.2012.08.015
Kruatrachue, B., Lewis, T.: Grain size determination for parallel processing. IEEE Softw. 5(1), 23–32 (1988). https://doi.org/10.1109/52.1991
Li, G., Chen, D., Wang, D., Zhang, D.: Task clustering and scheduling to multiprocessors with duplication, In: Proceedings International Parallel and Distributed Processing Symposium, pp. 8 (2003). https://doi.org/10.1109/IPDPS.2003.1213079
Pollack, J.B.: Recursive distributed representations. Artif. Intell. 46(1–2), 77–105 (1990). https://doi.org/10.1016/0004-3702(90)90005-K
Shirazi, B., Wang, M., Pathak, G.: Analysis and evaluation of heuristic methods for static task scheduling. J. Parallel Distrib. Comput. 10(3), 222–232 (1990). https://doi.org/10.1016/0743-7315(90)90014-G
Socher, R., Huang, E.H., Pennington, J., Ng, A.Y., Manning, C.D.: Dynamic pooling and unfolding recursive autoencoders for paraphrase detection, In: Proceedings of the 24th International Conference on Neural Information Processing Systems, Curran Associates Inc., Red Hook, NY, USA, pp. 801–809 (2011)
Socher, R., Perelygin, A., Wu, J., Chuang, J., Manning, C. D., Ng, A., Potts, C.: Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank, In: Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics, Seattle, Washington, USA, pp. 1631–1642 (2013)
Sperduti, A., Starita, A.: Supervised neural networks for the classification of structures. IEEE Trans. Neural Netw. 8(3), 714–735 (1997). https://doi.org/10.1109/72.572108
Yao, F., Pu, C., Zhang, Z.: Task Duplication-Based Scheduling Algorithm for Budget-Constrained Workflows in Cloud Computing. IEEE Access 9, 37262–37272 (2021). https://doi.org/10.1109/ACCESS.2021.3063456
Yeh-Ching C., Ranka, S.: Applications and performance analysis of a compile-time optimization approach for list scheduling algorithms on distributed memory multiprocessors, In: Proceedings of the ACM/IEEE Conference on Supercomputing, pp. 512–521 (1992). https://doi.org/10.1109/SUPERC.1992.236653
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Samandi, V., Tiňo, P., Bahsoon, R. (2022). Duplication Scheduling with Bottom-Up Top-Down Recursive Neural Network. In: Yin, H., Camacho, D., Tino, P. (eds) Intelligent Data Engineering and Automated Learning – IDEAL 2022. IDEAL 2022. Lecture Notes in Computer Science, vol 13756. Springer, Cham. https://doi.org/10.1007/978-3-031-21753-1_17
Download citation
DOI: https://doi.org/10.1007/978-3-031-21753-1_17
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-21752-4
Online ISBN: 978-3-031-21753-1
eBook Packages: Computer ScienceComputer Science (R0)