Abstract
Research is presented in this work to improve classification performance when using real-world training data for forecasting disease prediction likelihood. Optimisation techniques currently available are capable of providing highly efficient and accurate results however, performance potential can often be restricted when dealing with limited training resources. A novel approach is proposed with this work known as Synthetic Instance Model Optimisation (SIMO) which introduces Sequential Model-based Algorithm Configuration (SMAC) optimisation combined with Synthetic Minority Over-sampling Technique (SMOTE) for improving optimised prediction modelling. The SIMO approach generates additional synthetic instances from a limited training sample while simultaneously aiming to increase best algorithm performance. Results provided yield a partial solution for improving optimum algorithm performance when handling sparse training resources. Using the SIMO approach, noticeable performance accuracy and f-measure improvements were achieved over standalone SMAC optimisation. Results showed significant improvement when comparing collective training data with SIMO instance optimisation including individual performance accuracy increases of up to 46% and a mean overall increase for the entire 240 configurations of 13.96% over standard SMAC optimisation.
We would like to acknowledge The European Space Agency and The Academy of Opto-electronics, China as our research partnership in this work.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Bengio, Y., Goodfellow, I.J., Courville, A.: Optimization for training deep models. In: Deep Learning, pp. 238–290 (2015)
Chan, S., Treleaven, P., Capra, L.: Continuous hyperparameter optimization for large-scale recommender systems. In: Proceedings of 2013 IEEE International Conference on Big Data, Big Data 2013, pp. 350–358 (2013)
Chawla, N.V., Bowyer, K.W., Hall, L.O., Kegelmeyer, W.P.: SMOTE: synthetic minority over-sampling technique. J. Artif. Intell. Res. 16, 321–357 (2002)
Corne, D.W., Reynolds, A.P.: Optimisation and generalisation: footprints in instance space. In: Schaefer, R., Cotta, C., Kołodziej, J., Rudolph, G. (eds.) PPSN 2010. LNCS, vol. 6238, pp. 22–31. Springer, Heidelberg (2010). https://doi.org/10.1007/978-3-642-15844-5_3
Feurer, M., Springenberg, J.T., Hutter, F.: Initializing Bayesian hyperparameter optimization via meta-learning. In: Proceedings of the 29th Conference on Artificial Intelligence (AAAI 2015), pp. 1128–1135 (2015)
Fusco, T., Bi, Y., Wang, H., Browne, F.: Synthetic optimisation techniques for epidemic disease prediction modelling. In: Proceedings of the 7th International Conference on Data Science, Technology and Applications, pp. 95–106. SCITEPRESS - Science and Technology Publications (2018)
He, H., Bai, Y., Garcia, E.A., Li, S.: ADASYN: adaptive synthetic sampling approach for imbalanced learning. In: Proceedings of the International Joint Conference on Neural Networks, pp. 1322–1328 (2008)
Hutter, F., Hoos, H.H., Leyton-Brown, K.: Sequential model-based optimization for general algorithm configuration. In: Coello, C.A.C. (ed.) LION 2011. LNCS, vol. 6683, pp. 507–523. Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-25566-3_40
Hutter, F., Hoos, H.H., Leyton-Brown, K.: Parallel algorithm configuration. In: Hamadi, Y., Schoenauer, M. (eds.) LION 2012. LNCS, pp. 55–70. Springer, Heidelberg (2012). https://doi.org/10.1007/978-3-642-34413-8_5
Jian, C., Gao, J., Ao, Y.: A new sampling method for classifying imbalanced data based on support vector machine ensemble. Neurocomputing 193, 115–122 (2016)
Kotthoff, L., Thornton, C., Hoos, H.H., Hutter, F., Leyton-Brown, K.: Auto-WEKA 2.0: automatic model selection and hyperparameter optimization in WEKA. J. Mach. Learn. Res. 17, 1–5 (2016)
Krempl, G., Kottke, D., Lemaire, V.: Optimised probabilistic active learning (OPAL): for fast, non-myopic, cost-sensitive active classification. Mach. Learn. 100(2–3), 449–476 (2015)
Lin, Y.L., Hsieh, J.G., Wu, H.K., Jeng, J.H.: Three-parameter sequential minimal optimization for support vector machines. Neurocomputing 74(17), 3467–3475 (2011)
López, V., Triguero, I., Carmona, C.J., García, S., Herrera, F.: Addressing imbalanced classification with instance generation techniques: IPADE-ID. Neurocomputing 126, 15–28 (2014)
Sáez, J.A., Luengo, J., Stefanowski, J., Herrera, F.: SMOTE-IPF: addressing the noisy and borderline examples problem in imbalanced classification by a re-sampling method with filtering. Inf. Sci. 291(C), 184–203 (2015)
Settles, B.: Active Learning. Synth. Lect. Artif. Intell. Mach. Learn. 6(1), 1–114 (2012). http://www.morganclaypool.com/
Snoek, J., Larochelle, H., Adams, R.P.: Practical Bayesian optimization of machine learning algorithms. In: Advances in Neural Information Processing Systems 25, pp. 1–9 (2012)
Thornton, C., et al.: Auto-WEKA: combined selection and hyperparameter optimization of classification algorithms. In: Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 847–855 (2013)
Trebst, S., Troyer, M., Hansmann, U.H.E.: Optimized parallel tempering simulations of proteins. J. Chem. Phys. 124(17), 174903 (2006)
Acknowledgements
Training data is supplied by research partners at The European Space Agency in conjunction with The Academy of Opto-electronics, CAS, China.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Fusco, T., Bi, Y., Wang, H., Browne, F. (2019). Infectious Disease Prediction Modelling Using Synthetic Optimisation Approaches. In: Quix, C., Bernardino, J. (eds) Data Management Technologies and Applications. DATA 2018. Communications in Computer and Information Science, vol 862. Springer, Cham. https://doi.org/10.1007/978-3-030-26636-3_7
Download citation
DOI: https://doi.org/10.1007/978-3-030-26636-3_7
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-26635-6
Online ISBN: 978-3-030-26636-3
eBook Packages: Computer ScienceComputer Science (R0)