Abstract
Cross prediction is an active research area. Many research works have used cross prediction to predict the target system’s performance and power from the machine learning model trained on the source system. The source and target systems differ either in terms of instruction-set or hardware features. A widely used transfer learning technique utilizes the knowledge from a trained machine learning from one problem to predict targets in similar problems. In this work, we use transfer learning to achieve cross-system and cross-platform predictions. In cross-system prediction, we predict the physical system’s performance (runtime) and power from the simulation systems dataset while predicting performance and the power for target system from source system both having different instruction-set in cross-platform prediction. We achieve runtime prediction accuracy of 90% and 80% and power prediction accuracy of 75% and 80% in cross-system and cross-platform predictions, respectively, for the best performing deep neural network model. Furthermore, we have evaluated the accuracy of univariate and multivariate machine learning models, the accuracy of compute-intensive and data-intensive applications, and the accuracy of the simulation and physical systems.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Kawaguchi, S., Yachi, T.: Adaptive power efficiency control by computer power consumption prediction using performance counters. In: International Power Electronics Conference (IPEC-Hiroshima 2014 - ECCE ASIA), Hiroshima, pp. 3959–3965 (2014). https://doi.org/10.1109/TIA.2015.2466687
Lai, Z., Lam, K.T., Wang, C., Su, J.: PoweRock:power modeling and flexible dynamic power management for many-core architectures. IEEE Syst. J. 11(2), 600–612 (2017). https://doi.org/10.1109/JSYST.2015.2499307
Sîrbu, A., Babaoglu, O.: Predicting system-level power for a hybrid supercomputer. In: 2016 International Conference on High Performance Computing & Simulation (HPCS), Innsbruck, pp. 826–833 (2016)
Ma, J., Yan, G., Han, Y., Li, X.: An analytical framework for estimating scale-out and scale-up power efficiency of heterogeneous manycores. IEEE Trans. Comput. 65(2), 367–381 (2016)
Ardalani, N., Lestourgeon, C., Sankaralingam, K., Zhu, X.: Cross-architecture performance prediction (XAPP) using CPU code to predict GPU performance. In: Proceedings of the 48th International Symposium on Microarchitecture (MICRO-48). Association for Computing Machinery, New York, NY, USA, pp. 725–737 (2015)
Zheng, X., John, L.K., Gerstlauer, A.: LACross: learning-based analytical cross-platform performance and power prediction. Int. J. Parallel Prog. 45, 1488–1514 (2017)
Malakar, P., Balaprakash, P., Vishwanath, V., Morozov, V., Kumaran, K.: Benchmarking machine learning methods for performance modeling of scientific applications. In: IEEE/ACM Performance Modeling, Benchmarking and Simulation of High Performance Computer Systems (PMBS). Dallas, TX, USA, pp. 33–44 (2018)
Hoerl, Arthur E., Kennard, Robert W.: Ridge regression: biased estimation for nonorthogonal problems. Technometrics 12(1), 55–67 (1970)
Bishop, C.M.: Pattern Recognition and Machine Learning (Information Science and Statistics). Springer, Heidelberg (2006)
Rasmussen, C.E.: Gaussian processes in machine learning. In: Bousquet, O., von Luxburg, U., Rätsch, G. (eds.) Advanced Lectures on Machine Learning. Lecture Notes in Computer Science, vol. 3176. Springer, Heidelberg (2003)
Krzywinski, M., Altman, N.: Classification and regression trees. Nat. Methods 14, 757–758 (2017)
Breiman, L.: Random forests. Mach. Learn. 45, 5–32 (2001)
Geurts, P., Ernst, D., Wehenkel, L.: Extremely randomized trees. Mach. Learn. 63, 3–42 (2006)
Wang, H., Raj, B.: On the Origin of Deep Learning (2017). arXiv preprint arXiv:1702.07800
Devlin, J., et al.: Bert: Pre-training of Deep Bidirectional Transformers for Language Understanding (2018). arXiv preprint arXiv:1810.04805
Binkert, N., Beckmann, B., Black, G., Reinhardt, S.K., Saidi, A., Basu, A., Hestness, J., Hower, D.R., Krishna, T., Sardashti, S., Sen, R., Sewell, K., Shoaib, M., Vaish, N., Hill, M.D., Wood, D.A.: The gem5 simulator. SIGARCH Comput. Archit. News 39(2), 1–7 (2011)
Li, S., Ahn, J.H., Strong, R.D., Brockman, J.B., Tullsen, D.M., Jouppi, N.P: McPAT: an integrated power, area, and timing modeling framework for multicore and manycore architectures. In: Proceedings of the 42nd Annual IEEE/ACM International Symposium on Microarchitecture (MICRO 42). Association for Computing Machinery, New York, NY, USA, pp. 469–480 (2009)
Terpstra, D., Jagode, H., You, H., Dongarra, J.: Collecting performance data with PAPI-C. In: Tools for High Performance Computing 2009, Springer, Heidelberg, 3rd Parallel Tools Workshop, Dresden, Germany, pp. 157–173 (2010)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Mankodi, A., Bhatt, A., Chaudhury, B., Kumar, R., Amrutiya, A. (2021). Modeling Performance and Power on Disparate Platforms Using Transfer Learning with Machine Learning Models. In: Das, B., Patgiri, R., Bandyopadhyay, S., Balas, V.E. (eds) Modeling, Simulation and Optimization. Smart Innovation, Systems and Technologies, vol 206. Springer, Singapore. https://doi.org/10.1007/978-981-15-9829-6_18
Download citation
DOI: https://doi.org/10.1007/978-981-15-9829-6_18
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-15-9828-9
Online ISBN: 978-981-15-9829-6
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)