Abstract
Finding the optimal hyper-parameters values for a given problem is essential for most machine learning algorithms. In this paper, we propose a novel hyper-parameter optimization algorithm that is very simple to implement and still competitive with the state-of-the-art L-SHADE variant of Differential Evolution. While the most common method for hyper-parameter optimization is a combination of grid and manual search, random search has recently shown itself to be more effective and has been proposed as a baseline against which to measure other methods. In this paper, we compare three optimization algorithms, namely, the state-of-the-art L-SHADE algorithm, the random search algorithm, and our novel and simple adaptive random search algorithm. We find that our simple adaptive random search strategy is capable of finding parameters that achieve results comparable to the state-of-the-art L-SHADE algorithm, both of which achieve significantly better performance than random search when optimizing the hyper-parameters of the state-of-the-art XGBoost algorithm for 11 datasets. Because of the significant performance increase of our simple algorithm when compared to random search, we propose this as the new go-to method for tuning the hyper-parameters of machine learning algorithms when desiring a simple-to-implement algorithm, and also propose to use this algorithm as a new baseline against which other strategies should be measured.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Awad, N.H., Ali, M.Z., Suganthan, P.N., Reynolds, R.G.: An ensemble sinusoidal parameter adaptation incorporated with L-SHADE for solving CEC2014 benchmark problems. In: 2016 IEEE Congress on Evolutionary Computation (CEC), pp. 2958–2965. IEEE (2016)
Bergstra, J., Bengio, Y.: Random search for hyper-parameter optimization. J. Mach. Learn. Res. 13(Feb), 281–305 (2012)
Chen, T., Guestrin, C.: XGBoost: a scalable tree boosting system. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD 2016, pp. 785–794. ACM, New York (2016)
Guo, S.M., Tsai, J.S.H., Yang, C.C., Hsu, P.H.: A self-optimization approach for L-SHADE incorporated with eigenvector-based crossover and successful-parent-selecting framework on CEC 2015 benchmark set. In: 2015 IEEE Congress on Evolutionary Computation (CEC), pp. 1003–1010. IEEE (2015)
Jaderberg, M., et al.: Population based training of neural networks. arXiv preprint arXiv:1711.09846 (2017)
Klein, A., Falkner, S., Bartels, S., Hennig, P., Hutter, F.: Fast Bayesian optimization of machine learning hyperparameters on large datasets. arXiv preprint arXiv:1605.07079 (2016)
Lorenzo, P.R., Nalepa, J., Ramos, L.S., Pastor, J.R.: Hyper-parameter selection in deep neural networks using parallel particle swarm optimization. In: Proceedings of the Genetic and Evolutionary Computation Conference Companion, pp. 1864–1871. ACM (2017)
Loshchilov, I., Hutter, F.: CMA-ES for hyperparameter optimization of deep neural networks. arXiv preprint arXiv:1604.07269 (2016)
Martinez-Cantin, R.: BayesOpt: a Bayesian optimization library for nonlinear optimization, experimental design and bandits. J. Mach. Learn. Res. 15(1), 3735–3739 (2014)
Mohamed, A.W., Hadi, A.A., Fattouh, A.M., Jambi, K.M.: LSHADE with semi-parameter adaptation hybrid with CMA-ES for solving CEC 2017 benchmark problems. In: 2017 IEEE Congress on Evolutionary Computation (CEC), pp. 145–152. IEEE (2017)
Snoek, J., Larochelle, H., Adams, R.P.: Practical Bayesian optimization of machine learning algorithms. In: Advances in Neural Information Processing Systems, pp. 2951–2959 (2012)
Tanabe, R., Fukunaga, A.S.: Improving the search performance of SHADE using linear population size reduction. In: 2014 IEEE Congress on Evolutionary Computation (CEC), pp. 1658–1665. IEEE (2014)
Thornton, C., Hutter, F., Hoos, H.H., Leyton-Brown, K.: Auto-WEKA: automated selection and hyper-parameter optimization of classification algorithms. CoRR, abs/1208.3719 (2012)
Acknowledgement
This research was supported in part with computational resources at UIT provided by NOTUR, http://www.sigma2.no.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Geitle, M., Olsson, R. (2019). A New Baseline for Automated Hyper-Parameter Optimization. In: Nicosia, G., Pardalos, P., Umeton, R., Giuffrida, G., Sciacca, V. (eds) Machine Learning, Optimization, and Data Science. LOD 2019. Lecture Notes in Computer Science(), vol 11943. Springer, Cham. https://doi.org/10.1007/978-3-030-37599-7_43
Download citation
DOI: https://doi.org/10.1007/978-3-030-37599-7_43
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-37598-0
Online ISBN: 978-3-030-37599-7
eBook Packages: Computer ScienceComputer Science (R0)