Abstract
A PI is the range of values in which the real target value of a supervised learning task is expected to fall into, and it should combine two contrasting properties: to be as narrow as possible, and to include as many data observations as possible. This article presents an study on modelling Prediction Intervals (PI) with two Genetic Programming (GP) methods. The first proposed GP method is called CWC-GP, and it evolves simultaneously the lower and upper boundaries of the PI using a single fitness measure. This measure is the Coverage Width-based Criterion (CWC), which combines the width and the probability coverage of the PI. The second proposed GP method is called LUBE-GP, and it evolves independently the lower and upper boundaries of the PI. This method applies a multi-objective approach, in which one fitness aims to minimise the width and the other aims to maximise the probability coverage of the PI. Both methods were applied with the Direct and the Sequential approaches. In the former, the PI is assessed without the crisp prediction of the model. In the latter, the method makes use of the crisp prediction to find the PI boundaries. The proposed methods showed to have good potential on assessing PIs and the results pave the way to further investigations.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
Supplementary material available at https://bit.ly/3zsRfGP.
References
Kendall, A., Gal, Y.: What uncertainties do we need in bayesian deep learning for computer vision?. In: 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA (2017)
Deb, K., Pratap, A., Agarwal, S., Meyarivan, T.: A fast and elitist multiobjective genetic algorithm/ NSGA-II. IEEE Trans. Evol. Comput. 6(2), 182–197 (2002)
Vanneschi, L., Castelli, M., Scott, K., Popovic, A.: Accurate high performance concrete prediction with an alignment-based genetic programming system. Int. J. Concr. Struct. Mater. 12, 72 (2018)
Kiureghian, A.D., Ditlevsen, O.D.: Aleatoric or epistemic? Does it matter? Struct. Saf. 31(1), 105–112 (2009)
Cartagena, O., Parra, S., Muñoz-Carpintero, D., Marín, L.G., Sáez, D.: Review on fuzzy and neural prediction interval modelling for nonlinear dynamical systems. IEEE Access 9, 23357–2338 (2021)
Khosravi, A., Nahavandi, S., Creighton, D., Atiya, A.F.: Lower upper bound estimation method for construction of neural network-based prediction intervals. IEEE Trans. Neural Netw. 22(3), 337–346 (2011)
Taormina, R., Chau, K.: ANN-based interval forecasting of streamflow discharges using the LUBE method and MOFIPS. Eng. Appl. Artif. Intell. 45, 429–440 (2015)
Cruz, N., Marín, L.G., Sáez, D.: Neural network prediction interval based on joint supervision. In: 2018 International Joint Conference on Neural Networks (IJCNN) (2018)
Shrestha, D.L., Solomatine, D.P.: Machine learning approaches for estimation of prediction interval for the model output. Neural Netw. 19(2), 225–235 (2006)
Shrestha, D.L., Kayastha, N., Solomatine, D.P.: ANNs and other machine learning techniques in modelling models’ uncertainty. In: 19th International Conference, Limassol, Cyprus (2009)
Zhanga, C., Zhaoa, Y., Fanb, C., Lia, T., Zhanga, X., Lia, J.: A generic prediction interval estimation method for quantifying the T uncertainties in ultra-short-term building cooling load prediction. Appl. Therm. Eng. 173, 115261 (2020)
Khosravi, A., Nahavandi, S., Srinivasan, D., Khosravi, R.: Constructing optimal prediction intervals by using neural networks and bootstrap method. IEEE Trans. Neural Netw. Learn. Syst. 26(8), 1810–1815 (2015)
Blundell, C., Cornevise, J., Kavukcuoglu, K., Wierstra, D.: Weight uncertainty in neural networks. In: Proceedings of the 32nd International Conference on Machine Learning (ICML 2015), Lille (2015)
Lakshminarayanan, B., Pritzel, A., Blundell, C.: Simple and scalable predictive uncertainty estimation using deep ensembles. In: 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA (2017)
Maddox, W.J., Garipov, T., Izmailov, P., Vetrov, D., Wilson, A.G.: A simple baseline for bayesian uncertainty in deep learning. In: 33rd Conference on Neural Information Processing Systems (NIPS 2019), Vancouver, CA (2019)
Thuong, P.T., Hoai, N.X., Yao, X.: Combining conformal prediction and genetic programming for symbolic interval regression. In: Proceedings of the Genetic and Evolutionary Computation Conference (Berlin, Germany) (GECCO ’17), Berlin, DE (2017)
Koza, J.R.: Genetic programming: on the programming of computers by means of natural selection. Stat. Comput. 4, 87–112 (1992)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Brotto Rebuli, K., Giacobini, M., Tallone, N., Vanneschi, L. (2023). Single and Multi-objective Genetic Programming Methods for Prediction Intervals. In: De Stefano, C., Fontanella, F., Vanneschi, L. (eds) Artificial Life and Evolutionary Computation. WIVACE 2022. Communications in Computer and Information Science, vol 1780. Springer, Cham. https://doi.org/10.1007/978-3-031-31183-3_17
Download citation
DOI: https://doi.org/10.1007/978-3-031-31183-3_17
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-31182-6
Online ISBN: 978-3-031-31183-3
eBook Packages: Computer ScienceComputer Science (R0)