Skip to main content
Log in

Learning probability distributions in continuous evolutionary algorithms – a comparative review

  • Published:
Natural Computing Aims and scope Submit manuscript

Abstract

We present a comparative review of Evolutionary Algorithms that generate new population members by sampling a probability distributionconstructed during the optimization process. We present a unifying formulation for five such algorithms that enables us to characterize them based on the parametrization of the probability distribution, the learning methodology, and the use of historical information.The algorithms are evaluated on a number of test functions in order to assess their relative strengths and weaknesses. This comparative reviewhelps to identify areas of applicability for the algorithms and to guidefuture algorithmic developments.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  • Baluja S and Caruana R (1995) Removing the Genetics from the Standard Genetic Algorithm. Technical report, Computer Science Department, Carnegie Mellon University, Pittsburgh, PA. CMU-CS-95-141

    Google Scholar 

  • Beyer H-G and Deb K (2001) On self-adaptive features in real-parameter evolutionary algorithms. IEEE Transactions on Evolutionary Computation 5(3): 250–270

    Google Scholar 

  • Beyer H-G and Schwefel H-P (2002) Evolution strategies: A comprehensive introduction. Natural Computing 1(1), 3–52

    Google Scholar 

  • Bosman PAN and Thierens D (2000a) Expanding from discrete to continuous estimation of distribution algorithms: The IDEA. In: Schoenauer M, Deb K, Rudolph G, Yao X, Lutton E, Merelo JJ and Schwefel H-P (eds) Parallel Problem Solving from Nature — PPSN VI, pp. 767–776. Springer, Berlin

    Google Scholar 

  • Bosman PAN and Thierens D (2000b) Mixed ID{ie111-1}As. Technical report, Utrech University. UU-CS-2000-45

  • Bosman PAN and Thierens D (2001) Advancing continuous IDEAs with mixture distributions and factorization selection metrics. In: Optimization by Building and Using Probabilistic Models (OBUPM) 2001, pp. 208–212. San Francisco, California, USA

  • Hansen N, Müller SD and Koumoutsakos P (2003) Reducing the time complexity of the derandomized evolution strategy with covariance matrix adaptation (CMA-ES). Evolutionary Computation 11(1): 1–18

    Google Scholar 

  • Hansen N and Ostermeier A (1996) Adapting arbitrary normal mutation distributions in evolution strategies: The covariance matrix adaptation. In: Proceedings of the 1996 IEEE Conference on Evolutionary Computation (ICEC '96), pp. 312–317

  • Hansen N and Ostermeier A (2001) Completely derandomized self-adaptation in evolution strategies. Evolutionary Computation 9(2): 159–195

    Google Scholar 

  • Larrañaga P (2002) A review on estimation of distribution algorithms. In: Larrañaga P and Lozano JA (eds) Estimation of Distribution Algorithms, pp. 80–90. Kluwer Academic Publishers, Dordrecht

    Google Scholar 

  • Mühlenbein H and Paass G (1996) From recombination of genes to the estimation of distributions: I. binary parameters. Lecture Notes in Computer Science 1141: 178–187

  • Ocenasek J and Schwarz J (2002) Estimation of distribution algorithm for mixed continuous — discrete optimization problems. In: 2nd Euro-International Symposium on Computational Intelligance, pp. 227–232. IOS Press, Kosice, Slowakia

    Google Scholar 

  • Ostermeier A, Gawelczyk A and Hansen N (1994) Step-size adaptation based on non-local use of selection information. In: Davidor Y, Schwefel H-P and Männer R (eds) Parallel Problem Solving from Nature — PPSN IV, Proceedings, Jerusalem, pp. 189–198. Springer, Berlin

    Google Scholar 

  • Parzen E (1962) On estimation of probability density function and mode. Annual Mathematical Statistics 33: 1065–1076

    Google Scholar 

  • Pelikan M and Goldberg DE (2001) Escaping hierarchical traps with competent genetic algorithms. IlliGAL Report No. 2001003, Illinois Genetic Algorithms Laboratory, University of Illinois at Urbana-Champaign, Urbana, IL

    Google Scholar 

  • Pelikan M, Goldberg DE and Cantú-Paz E (1999) BOA: The Bayesian optimization algorithm. In: Banzhaf W, Daida J, Eiben AE, Garzon MH, Honavar V, Jakiela M and Smith RE (eds) Proceedings of the Genetic and Evolutionary Computation Conference GECCO-99, Vol. I. Orlando, FL, pp. 525–532. Morgan Kaufmann Publishers, San Francisco, CA

    Google Scholar 

  • Pelikan M, Goldberg DE and Sastry K (2000) Bayesian optimization algorithm, decision graphs, and Occam's razor. IlliGAL Report No. 2000020, Illinois Genetic Algorithms Laboratory, University of Illinois at Urbana-Champaign, Urbana, IL

    Google Scholar 

  • Rechenberg I (1973) Evolutionsstrategie: Optimierung technischer Systeme nach Prinzipien der biologischen Evolution. Fromman-Holzboog Verlag, Stuttgart

    Google Scholar 

  • Schwefel H-P (1995) Evolution and Optimum Seeking. John Wiley & Sons, New York

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Petros Koumoutsakos.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Kern, S., Müller, S.D., Hansen, N. et al. Learning probability distributions in continuous evolutionary algorithms – a comparative review. Natural Computing 3, 77–112 (2004). https://doi.org/10.1023/B:NACO.0000023416.59689.4e

Download citation

  • Issue Date:

  • DOI: https://doi.org/10.1023/B:NACO.0000023416.59689.4e

Navigation