Abstract
Too often, when comparing a set of optimization algorithms, little effort, if any at all, is spent for finding the parameter settings which let them perform at their best on a given optimization task. Within this context, automatizing the choice of their parameter settings can be seen as a way to perform fair comparisons between optimization algorithms.
In this paper we first compare the performances of two standard PSO versions using the “standard” parameters suggested in the literature. Then, we automatically tune the parameter values of both algorithms using a meta-optimization environment, to allow the two versions to perform at their best.
As expected, results obtained by the optimized version are substantially better than those obtained with the standard settings. Moreover, they generalize well on other functions, allowing one to draw interesting conclusions regarding the PSO parameter settings that are commonly used in the literature.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Johnson, D.S.: A theoretician’s guide to the experimental analysis of algorithms. In: Goldwasser, M.H., Johnson, D.S., McGeoch, C.C. (eds.) Data Structures, Near Neighbor Searches, and Methodology: Fifth and Sixth DIMACS Implementation Challenges, vol. 59, pp. 215–250. American Mathematical Society, Providence (2002)
Kennedy, J., Eberhart, R.: Particle swarm optimization. In: IEEE International Conference on Neural Networks, ICNN 1995, vol. 4, pp. 1942–1948 (1995)
Poli, R.: Analysis of the publications on the applications of Particle Swarm Optimisation. J. Artif. Evol. Appl. 2008, 1–10 (2008)
Clerc, M.: Particle Swarm Optimization. Wiley, New York (2010)
Vazquez, J., Valdez, F., Melin, P.: Comparative study of social network structures in PSO. In: Castillo, O., Melin, P., Pedrycz, W., Kacprzyk, J. (eds.) Recent Advances on Hybrid Approaches for Designing Intelligent Systems. Studies in Computational Intelligence, vol. 547, pp. 239–254. Springer, Basel (2014)
Banks, A., Vincent, J., Anyakoha, C.: A review of particle swarm optimization. part I: background and development. Nat. Comput. 6(4), 467–484 (2007)
Clerc, M.: Standard particle swarm optimisation from 2006 to 2011. Technical report, Particle Swarm Central (2011)
Jiang, M., Luo, Y., Yang, S.: Stochastic convergence analysis and parameter selection of the standard particle swarm optimization algorithm. Inf. Process. Lett. 102(1), 8–16 (2007)
Bonyadi, M.R., Michalewicz, Z.: SPSO 2011: analysis of stability; local convergence; and rotation sensitivity. In: Proceedings of the 2014 Conference on Genetic and Evolutionary Computation. GECCO 2014, pp. 9–16. ACM (2014)
Zambrano-Bigiarini, M., Clerc, M., Rojas, R.: Standard particle swarm optimisation 2011 at CEC-2013: a baseline for future pso improvements. In: IEEE Congress on Evolutionary Computation, CEC 2013, pp. 2337–2344, June 2013
Liang, J., Qu, B., Suganthan, P.: Problem definitions and evaluation criteria for the CEC 2013 special session on real-parameter optimization. Technical report 201212, Computational Intelligence Laboratory, Zhengzhou University (2013)
Mercer, R., Sampson, J.: Adaptive search using a reproductive metaplan. Kybernetes 7, 215–228 (1978)
Grefenstette, J.: Optimization of control parameters for genetic algorithms. IEEE Trans. Syst. Man Cybern. 16(1), 122–128 (1986)
Nannen, V., Eiben, A.E.: Relevance estimation and value calibration of evolutionary algorithm parameters. In: International Joint Conference on Artifical Intelligence, IJCAI 2007, pp. 975–980 (2007)
Larrañaga, P., Lozano, J.A.: Estimation of Distribution Algorithms: A New Tool for Evolutionary Computation. Kluwer Academic Publishers, Norwell (2001)
Meissner, M., Schmuker, M., Schneider, G.: Optimized Particle Swarm Optimization (OPSO) and its application to artificial neural network training. BMC Bioinform. 7, 125 (2006)
Pedersen, M.E.H.: Tuning and simplifying heuristical optimization. Master’s thesis, University of Southampton (2010)
Hutter, F., Hoos, H.H., Leyton-Brown, K., Stützle, T.: ParamILS: an automatic algorithm configuration framework. J. Artif. Intell. Res. 36(1), 267–306 (2009)
Hutter, F., Hoos, H.H., Stützle, T.: Automatic algorithm configuration based on local search. In: National Conference on Artificial Intelligence, AAAI 2007, vol. 7, pp. 1152–1157 (2007)
Eiben, A.E., Smit, S.K.: Parameter tuning for configuring and analyzing evolutionary algorithms. Swarm Evol. Comput. 1(1), 19–31 (2011)
Ugolotti, R., Nashed, Y.S.G., Mesejo, P., Cagnoni, S.: Algorithm configuration using GPU-based metaheuristics. In: Genetic and Evolutionary Computation Conference (GECCO) Companion Proceedings, pp. 221–222. ACM (2013)
Storn, R., Price, K.: Differential evolution- a simple and efficient adaptive scheme for global optimization over continuous spaces. Technical report, International Computer Science Institute (1995)
Loshchilov, I., Stützle, T., Liao, T.: Ranking results of CEC 2013 special session & competition on real-parameter single objective optimization
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing Switzerland
About this paper
Cite this paper
Ugolotti, R., Cagnoni, S. (2016). A Fair Comparison Between Standard PSO Versions. In: Rossi, F., Mavelli, F., Stano, P., Caivano, D. (eds) Advances in Artificial Life, Evolutionary Computation and Systems Chemistry. WIVACE 2015. Communications in Computer and Information Science, vol 587. Springer, Cham. https://doi.org/10.1007/978-3-319-32695-5_1
Download citation
DOI: https://doi.org/10.1007/978-3-319-32695-5_1
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-32694-8
Online ISBN: 978-3-319-32695-5
eBook Packages: Computer ScienceComputer Science (R0)