Skip to main content

Advertisement

Log in

Problem solution sustenance in XCS: Markov chain analysis of niche support distributions and the impact on computational complexity

  • Published:
Genetic Programming and Evolvable Machines Aims and scope Submit manuscript

Abstract

Michigan-style learning classifier systems iteratively evolve a distributed solution to a problem in the form of potentially overlapping subsolutions. Each problem niche is covered by subsolutions that are represented by a set of predictive rules, termed classifiers. The genetic algorithm is designed to evolve classifier structures that together cover the whole problem space and represent a complete problem solution. An obvious challenge for such an online evolving, distributed knowledge representation is to continuously sustain all problem subsolutions covering all problem niches, that is, to ensure niche support. Effective niche support depends both on the probability of reproduction and on the probability of deletion of classifiers in a niche. In XCS, reproduction is occurrence-based whereas deletion is support-based. In combination, niche support is assured effectively. In this paper we present a Markov chain analysis of the niche support in XCS, which we validate experimentally. Evaluations in diverse Boolean function settings, which require non-overlapping and overlapping solution structures, support the theoretical derivations. We also consider the effects of mutation and crossover on niche support. With respect to computational complexity, the paper shows that XCS is able to maintain (partially overlapping) niches with a computational effort that is linear in the inverse of the niche occurrence frequency.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

Notes

  1. All parameters used in the paper are listed in Appendix A.

  2. We remind the reader that given a random variable X the cumulative distribution function \(F(x)\) is defined as \(F(x)=P(X\leq x)\).

  3. The same analysis on the even smaller 6-multiplexer is available in [14].

References

  1. E. Bernadó, X. Llorà, and J. M. Garrell, “XCS and GALE: A comparative study of two learning classifier systems and six other learning algorithms on classification tasks,” in Lanzi et al. [42], pp. 115–132.

  2. E. Bernadó-Mansilla and J. M. Garrell-Guiu, “Accuracy-based learning classifier systems: Models, analysis, and applications to classification tasks,” Evolutionary Computation, vol. 11, pp. 209–238, 2003.

  3. H.-G. Beyer, U.-M. O’Reily, D. V. Arnold, W. Banzhaf, C. Blum, E. W. Bonabeau, E. Cant-Paz, D. Dasgupta, K. Deb, J. A. Foster, E. D. de Jong, H. Lipson, X. Llora, S. Mancoridis, M. Pelikan, G. R. Raidl, T. Soule, A. M. Tyrrell, J.-P. Watson, and E. Zitzler (eds.). GECCO 2005: Genetic and Evolutionary Computation Conference: Volume 2. Association for Computing Machinery, Inc: New York, NY, 2005.

  4. C. L. Bridges and D. E. Goldberg, “An analysis of reproduction and crossover in a binary-coded genetic algorithm,” in Grefenstette [26], pp. 9–13.

  5. L. Bull, “Simple Markov models of the genetic algorithm in classifier systems: Accuracy-based fitness,” in Lanzi et al. [41], pp. 21–28.

  6. L. Bull, “Simple Markov models of the genetic algorithm in classifier systems: Multi-step sasks,” in Lanzi et al. [41], pp. 29–36.

  7. L. Bull and J. Hurst, “Self-adaptive mutation in ZCS controllers,” in Proceedings of the EvoNet Workshops - EvoRob 2000, S. Cagnoni, R. Poli, G. Smith, D. Corne, M. Oates, E. Hart, P.-L. Lanzi, E. Willem, Y. Li, B. Paechter, and T. C. Fogarty (eds.), Springer-Verlag: Berlin Heidelberg, 2000, pp. 339–346.

  8. L. Bull and J. Hurst, “ZCS Redux,” Evolutionary Computation, vol. 10, pp. 185–205, 2003.

  9. L. Bull, J. Hurst and A. Tomlinson, “Self-adaptive mutation in classifier system controllers,” in From Animals to Animats 6: Proceedings of the Sixth International Conference on Simulation of Adaptive Behavior, J.-A. Meyer, A. Berthoz, D. Floreano, H. Roitblat, and S. W. Wilson (eds.), MIT Press: Cambridge, MA, 2000, pp. 460–467.

  10. M. V. Butz, “Documentation of XCS + TS C-Code 1.2,” Technical Report 2003023, Illinois Genetic Algorithms Laboratory – University of Illinois at Urbana-Champaign, 2003.

  11. M. V. Butz, “Kernel-based, ellipsoidal conditions in the real-valued XCS classifier system,” in Beyer et al. [3], pp 1835–1842.

  12. M. V. Butz, in Rule-Based Evolutionary Online Learning Systems: A Principled Approach to LCS Analysis and Design. Studies in Fuzziness and Soft Computing. Springer-Verlag: Berlin Heidelberg, 2006.

  13. M. V. Butz, D. E. Goldberg, and P. L. Lanzi, “Bounding learning time in XCS,” in Genetic and Evolutionary Computation Conference - GECCO 2004: Part II, R. Poli, W. Banzhaf, H.-G. Beyer, E. Burke, P. Darwen, D. Dasgupta, D. Floreano, J. Foster, M. Harman, P. L. Lanzi, L. Spector, A. G. B. Tettamanzi, D. Thierens and A. M. Tyrrell (eds.), Springer-Verlag: Berlin Heidelberg, 2004, pp. 739–750.

  14. M. V. Butz, D. E. Goldberg, P. L. Lanzi, and K. Sastry, “Bounding the population size to ensure niche support in XCS,” Technical Report 2004033, Illinois Genetic Algorithms Laboratory – University of Illinois at Urbana-Champaign, 2004.

  15. M. V. Butz, D. E. Goldberg, and K. Tharakunnel, “Analysis and improvement of fitness exploitation in XCS: Bounding models, tournament selection, and bilateral accuracy,” Evolutionary Computation, vol. 11, pp. 239–277, 2003.

  16. M. V. Butz, T. Kovacs, P. L. Lanzi, and S. W. Wilson,“Toward a theory of generalization and learning in XCS,” IEEE Transactions on Evolutionary Computation, vol. 8, pp. 28–46, 2004.

  17. M. V. Butz, K. Sastry, and D. E. Goldberg, “Strong, stable, and reliable fitness pressure in XCS due to tournament selection,” Genetic Programming and Evolvable Machines, vol. 6, pp. 53–77, 2004.

  18. M. V. Butz and S. W. Wilson, “An algorithmic description of XCS,” Journal of Soft Computing, vol. 6 pp. 144–153, 2002.

  19. W. J. Conover, in Practical Nonparametric Statistics, 3rd edition, John Wiley & Sons: New York, NY, USA, 1998.

  20. K. A. De Jong, “An analysis of the behavior of a class of genetic adaptive systems,” PhD thesis, University of Michigan, Ann Arbor, 1975. University Microfilms No. 76-9381.

  21. P. W. Dixon, D. W. Corne, and M. J. Oates, “A preliminary investigation of modified XCS as a generic data mining tool,” in Lanzi et al. [42], pp. 133–150.

  22. R. A. Fisher. Statistical Methods for Research Workers. Oliver and Boyd, Edinburgh, GB, 1925. Available online at \( {\tt http://psychclassics.yorku.ca/Fisher/Methods/}\).

  23. M. Friendly, Visualizing Categorical Data. SAS Institute: Cary, NC, 2000.

  24. D. Goldberg and P. Segrest, “Finite Markov chain analysis of genetic algorithms,” in Grefenstette [26], pp. 1–8.

  25. D. E. Goldberg, “Simple genetic algorithms and the minimal, deceptive problem,” in Genetic Algorithms and Simulated Annealing, L. Davis (eds.), Research Notes in Artificial Intelligence, Pitman, London, 1987, pp. 74–88.

  26. J. J. Grefenstette (ed.), Proceedings of the Second International Conference on Genetic Algorithms, Lawrence Erlbaum Associates, Inc: Mahwah, NJ, USA, 1987.

  27. G. Harik, “Finding multiple solutions in problems of bounded difficulty,” in Proceedings of the Sixth International Conference on Genetic Algorithms, L. Eschelman (eds.), Morgan Kaufmann: San Francisco, CA, 1995, pp. 24–31,

  28. G. Harik, E. Cantú-Paz, D. E. Goldberg, and B. L. Miller, “The gambler's ruin problem, genetic algorithms, and the sizing of populations,” Evolutionary Computation, vol. 7, pp. 231–253, 1999.

  29. J. H. Holland, Adaptation in Natural and Artificial Systems, 2nd edition, 1992, University of Michigan Press: Ann Arbor, MI, 1975.

  30. J. H. Holland, Hidden Order: How Adaptation Builds Complexity. Perseus Books: Cambridge, MA, USA, 1995.

  31. J. Horn, “Finite Markov chain analysis of genetic algorithms with niching,” in Proceedings of the Fifth International Conference on Genetic Algorithms, S. Forrest (eds.), Morgan Kaufmann: San Francisco, CA, 1993, pp. 110–117,

  32. J. Horn, D. E. Goldberg, and K. Deb, “Implicit niching in a learning classifier system: Nature's way,“ Evolutionary Computation, vol. 2, pp. 37–66, 1994.

  33. J. Hurst and L. Bull, “A Self-Adaptive XCS,” in Lanzi et al. [42], pp. 57–73.

  34. K. A. D. Jong and W. M. Spears, “Learning concept classification rules using genetic algorithms,” in Proceedings of the Twelfth International Conference on Artificial Intelligence IJCAI-91, J. Mylopoulos and R. Reiter (eds.), vol. 2, Morgan Kaufmann: San Francisco, CA, 1991.

  35. L. Kleinroch, Queueing Systems: Theory, John Wiley & Sons: New York, NY, USA, 1975.

  36. T. Kovacs, “XCS classifier system reliably evolves accurate, complete, and minimal representations for boolean functions,” in Soft Computing in Engineering Design and Manufacturing, P. K. Chawdhry, R. Roy and R. K. Pant (eds.), Springer-Verlag: New York, NY, USA, 1997, pp. 59–68.

  37. T. Kovacs, “Deletion schemes for classifier systems,” in Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-99), W. Banzhaf, J. Daida, A. E. Eiben, M. H. Garzon, V. Honavar, M. Jakiela and R. E. Smith (eds.), Morgan Kaufmann: San Francisco, CA, 1999, pp. 329–336.

  38. T. Kovacs, Strength or Accuracy: Credit Assignment in Learning Classifier Systems, Springer-Verlag: Berlin Heidelberg, 2003.

  39. P. L. Lanzi, “The XCS library,” \( {\tt http://xcslib.sourceforge.net, 2002}\).

  40. P. L. Lanzi, D. Loiacono, S. W. Wilson and D. E. Goldberg, “Extending XCSF beyond linear approximation,” in Beyer et al. [3], pp. 1827–1834.

  41. P. L. Lanzi, W. Stolzmann, and S. W. Wilson (eds.), Advances in Learning Classifier Systems: Third International Workshop, IWLCS 2000: Paris, France, September 2000: Revised Papers, volume 1996 of LNAI, Springer-Verlag: Berlin Heidelberg, 2001.

  42. P. L. Lanzi, W. Stolzmann, and S. W. Wilson (eds.), Advances in Learning Classifier Systems: 4th International Workshop, IWLCS 2001: San Francisco, CA, USA, July 2001: Revised Papers, volume 2321 of LNAI, Springer-Verlag: Berlin Heidelberg, 2002.

  43. D. G. Luenberger, Introduction to Dynamic Systems: Theory, Models, and Applications, John Wiley & Sons: New York, NY, USA, May 1979.

  44. S. W. Mahfoud, “Crowding and preselection revisited,” in Parallel Problem Solving from Nature, 2, R. Männer and B. Manderick (eds.), Elsevier: Amsterdam, 1992, pp. 27–36.

  45. R. Development Core Team. R: A language and environment for statistical computing. R Foundation for Statistical Computing, Vienna, Austria, 2004. ISBN 3-900051-00-3.

  46. G. Venturini, “Adaptation in dynamic environments through a minimal probability of exploration,” in From Animals to Animats 3: Proceedings of the Third International Conference on Simulation of Adaptive Behavior, D. Cliff, P. Husbands, J.-A. Meyer, and S. W. Wilson (eds.), MIT Press: Cambridge, MA, 1994, pp. 371–381.

  47. S. W. Wilson, “Classifier systems and the animat problem,” Machine Learning, vol. 2, pp. 199–228, 1987.

  48. S. W. Wilson, “ZCS: A zeroth level classifier system,” Evolutionary Computation, vol. 2, pp. 1–18, 1994.

  49. S. W. Wilson, “Classifier fitness based on accuracy,” Evolutionary Computation, vol. 3, pp. 149–175, 1995.

  50. S. W. Wilson, “Generalization in the XCS classifier system,” in Genetic Programming 1998: Proceedings of the Third Annual Conference, J. R. Koza, W. Banzhaf, K. Chellapilla, K. Deb, M. Dorigo, D. B. Fogel, M. H. Garzon, D. E. Goldberg, H. Iba, and R. Riolo (eds.), Morgan Kaufmann: San Francisco, CA, 1998, pp. 665–674.

Download references

Acknowledgments

We are grateful to Xavier Llorà, Kei Onishi, Martin Pelikan, Stewart Wilson, Tian-Li Yu, and the whole IlliGAL lab for their help and the useful discussions.

This work was supported by the Air Force Office of Scientific Research, Air Force Materiel Command, USAF, under grant FA9550-06-1-0096, the National Science Foundation under ITR grant DMR-03-25939 (at Materials Computation Center, UIUC), The US Government is authorized to reproduce and distribute reprints for Government purposes notwithstanding any copyright notation thereon. Butz's contribution received additional funding from the European commission contract no. FP6-511931.

The views and conclusions contained herein are those of the authors and should not be interpreted as necessarily representing the official policies or endorsements, either expressed or implied, of any of the organizations mentioned above

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Martin V. Butz.

Additional information

Communicated by: W. Banzhaf

Appendices

Appendix A. List of symbols

Table 8
Table 9

Appendix B. Closed-form equation

In this section, we develop the closed form equation for the Markov chain in Fig. 1. We start from the following equation, introduced in Section 4:

$$u_k = r_{k-1}u_{k-1} + s_k u_k + d_{k+1}u_{k+1}\label{eq:fp}$$
(13)

by replacing the probabilities \(r_k, d_k\), and \(s_k\), with their actual values (see Section 4 for details):

$$ \begin{array}{l}r_k = p\left(1-\displaystyle\frac{k}{N}\right)\\ s_k = \left(1-p\right)\left(1-\displaystyle\frac{k}{N}\right) + p\displaystyle\frac{k}{N}\\ d_k = \left(1-p\right)\displaystyle\frac{k}{N}\end{array} $$

we obtain,

$$\left[ p\left(1-\frac{k}{N}\right) + \frac{k}{N}\left(1-p \right) \right] u_k =\left(1-p \right)\left( \frac{k+1}{N} \right)u_{k+1} +p\left( 1- \frac{k-1}{N}\right)u_{k-1}$$

finally, by dividing by \((1-p)u_{k-1}\), we derive the following equation:

$$\left[ \frac{p}{1-p}(N-k) + k\right]\frac{u_k}{u_{k-1}} = \left(k+1\right)\frac{u_{k+1}}{u_{k-1}} + \frac{p}{1-p}(N-k+1)\label{eq:fixedpoint2}$$
(14)

To derive a closed-form solution for probability \(u_k\) we first use Eq. (14) to derive a closed-form equation for the ratio \(\frac{u_{k}}{u_{0}}\). Next, we use the equation for \(\frac{u_k}{u_0}\) and the condition \(\sum_{k=0}^{N}u_k = 1\), to derive the closed-form solution for \(u_k\).

B.1. Closed-form equation for \(u_k/u_0\)

As the very first step, we write the following fixed point equation for the transitions between state 0 and state 1:

$$u_0 = s_0 u_0+d_1u_1.$$
(15)

By substituting the values of \(s_0\) and \(d_1\) we obtain:

$$ \begin{array}{l}u_0 = (1-p)u_0+(1-p)\displaystyle\frac{1}{N}u_1,\\ pu_0 = (1-p)\displaystyle\frac{1}{N}u_1,\end{array} $$

from which we derive equation:

$$\frac{u_1}{u_0} = \frac{p}{1-p}N.\label{eq:caseu1}$$
(16)

To derive the equation for \(u_2/u_0\) we start from Eq. (14) and set \(k=1\):

$$\left[ \frac{p}{1-p}(N-1) + 1\right]\frac{u_1}{u_0} = 2\frac{u_2}{u_0} + \frac{p}{1-p}N,$$

so that

$$\frac{u_2}{u_0} = \frac{1}{2}\left[\left(\frac{p}{1-p}(N-1) + 1\right)\frac{u_1}{u_0} - \frac{p}{1-p}N\right].$$

We now replace \(u_1/u_0\) with Eq. (16) :

$$\displaylines{ \displaystyle\frac{u_2}{u_0} = \displaystyle\frac{1}{2}\left[\left(\frac{p}{1-p}(N-1) + 1\right)\frac{u_1}{u_0} - \frac{p}{1-p}N\right] \cr = \displaystyle\frac{1}{2}\left[\left(\frac{p}{1-p}(N-1) + 1\right)\frac{p}{1-p}N - \frac{p}{1-p}N\right]\cr = \displaystyle\frac{1}{2}\left[\frac{p}{1-p}(N-1)\frac{p}{1-p}N\right]\cr = \displaystyle\frac{N(N-1)}{2}\left(\frac{p}{1-p}\right)^2\cr = {N\choose 2}\left(\displaystyle\frac{p}{1-p}\right)^2 \label{eq:caseu2} }$$
(17)

This leads us to the hypothesis that

$$ \frac{u_k}{u_0}=\left(\begin{array}{c} N\\ k\\ \end{array}\right) \left(\frac{p}{1-p}\right)^k, $$
(18)

which we prove by induction. Using Eq. (18), we can first derive that:

$$\displaylines{ u_{k+1} = \frac{N-k}{k+1}\frac{p}{1-p}u_k\label{eq:ukp1}\cr u_k = \frac{N-k+1}{k}\frac{p}{1-p}u_{k-1}\label{eq:uk}\cr u_{k-1} = \frac{1-p}{p}\frac{k}{N-k+1}u_k\label{eq:ukm1} }$$
(19)

With Eq. (14) substituting Eq. (21) as the inductive step, we now derive

$$\displaylines{ u_{k+1} = \frac{\big(\big(\frac{p}{1-p}(N-k) + k\big)\frac{u_k}{u_{k-1}} - \frac{p}{1-p} (N-k+1)\big)u_{k-1}}{k+1}\cr = \frac{\big(\frac{p}{1-p}\big)^2\frac{(N-k)(N-k+1)}{k}u_{k-1}}{k+1}\cr = \frac{N-k}{k+1}\frac{p}{1-p}u_k, }$$
(22)

which proves the hypothesis.

B.2. Derivation of \(u_0\)

To derive a closed-form for \(u_k\) from the equation for \(u_k/u_0\), we use the subsidiary condition:

$$\sum_{k=0}^{N}u_k=1\label{eq:sub}$$
(25)

We divide both terms by \(u_0\)

$$\sum_{k=0}^{N}\frac{u_k}{u_0} = \frac{1}{u_0},$$
(26)

where

$$\displaylines{ \sum_{k=0}^{N}\frac{u_k}{u_0} = \sum_{k=0}^{N}{N \choose k}\left(\frac{p}{1-p}\right)^k \cr = \left[\sum_{k=0}^{N}{N \choose k}p^k(1-p)^{N-k}\right]\frac{1}{(1-p)^N}, }$$

where the term “ \(\sum_{k=0}^{N}{N\choose k}p^k(1-p)^{N-k}\)” is equal to “ \(\left[p+(1-p)\right]^N\)”, that is 1, so that:

$$\sum_{k=0}^{N}\frac{u_k}{u_0}= \frac{1}{(1-p)^N}$$
(27)

and accordingly,

$$u_0 = (1-p)^N.\label{eq:u0}$$
(28)

B.3. Closed-form equation for \(u_k\)

By combining Eqs. (18) and (28), we derive the closed-form equation for \(u_k\) as follows:

$$ \begin{array}{lll}u_k & = & \displaystyle{N\choose k}\left(\displaystyle\frac{p}{1-p}\right)^ku_0\\ & = & \displaystyle{N\choose k}\left(\displaystyle\frac{p}{1-p}\right)^k(1-p)^N\\ & = & \displaystyle{N\choose k}p^k(1-p)^{N-k}\end{array} $$

Note that the same derivation is possible noting that the proposed Markov chain results in an Engset distribution [35].

Rights and permissions

Reprints and permissions

About this article

Cite this article

Butz, M.V., Goldberg, D.E., Lanzi, P.L. et al. Problem solution sustenance in XCS: Markov chain analysis of niche support distributions and the impact on computational complexity. Genet Program Evolvable Mach 8, 5–37 (2007). https://doi.org/10.1007/s10710-006-9012-8

Download citation

  • Received:

  • Revised:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10710-006-9012-8

Keywords

Navigation