Abstract
A new technique for improving the classification performance of learning classifier systems (LCS) was developed and applied to a real-world data mining problem. EpiCS, a stimulus-response LCS, was adapted to perform prevalence-based bootstrapping, wherein data from training and testing sets were sampled according to the prevalence of the individual classes, rather than randomly using the class distribution inherent in the data. Prevalence-based bootstrapping was shown to improve classification performance significantly on training and testing. Furthermore, this procedure was shown to enhance EpiCS’s classification performance on testing compared to a well-known decision tree inducer (C4.5) when similar bootstrapping procedures were applied to the latter.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Abe, N. and Mamitsuka, H.: Query learning strategies using boosting and bagging. In: Shavlik, J. (ed.): Machine Learning. Proceedings of the Fifteenth International Conference (ICML’98). San Francisco, Morgan Kaufmann Publishers (1998) 1–9.
Association for the Advancement of Automotive Medicine: The Abbreviated Injury Scale, 1990 Revision. Des Plaines, IL (1990).
Bauer, E. and Kohavi, R.: An empirical comparison of voting classification algorithms bagging, boosting, and variants. Machine Learning 36 (1999) 105–139.
Bonelli, P., Parodi, A., Sen, S., and Wilson, S.: NEWBOOLE: A fast GBML system, in: Porter, B. and Mooney, R. (eds.), Machine Learning: Proceedings of the Seventh International Conference. Morgan Kaufmann, San Mateo, CA (1990) 153–159.
Efron, B. and Tibshirani, R.J.: An Introduction to the Bootstrap. Chapman and Hall, New York (1993).
Goldberg, D.E.: Genetic Algorithms in Search, Optimization, and Machine Learning. Addison-Wesley, New York (1989).
Harries, M.: Boosting a strong learner: evidence against the minimum margin. In: Bratko, I. and Dzeroski, S. (eds.): Machine Learning. Proceedings of the Sixteenth International Conference (ICML’ 99). Morgan Kaufmann Publishers, San Francisco (1999) 171–180.
Holland, J.H., Holyoak, K.J., Nisbett, R.E., and Thagard, P.R.: Induction: Processes of Inference, Learning, and Discovery. The MIT Press, Cambridge, MA (1986).
Holmes, J.H.: A genetics-based machine learning approach to knowledge discovery in clinical data, Journal of the American Medical Informatics Association Suppl (1996) 883.
Holmes, J.H.: Discovery of Disease Risk with a Learning Classifier System, in: Baeck, T. (ed.): Proceedings of the Seventh International Conference on Genetic Algorithms (SanFrancisco, Morgan Kaufmann (1997) 426–433.
Holmes, J.H., Winston, F.K., Durbin, D.R., et al: The Partners for Child Passenger Safety Project: An information infrastructure for injury surveillance, Journal of the American Medical Informatics Association Suppl (1998) 1016.
Holmes J.H.: Differential negative reinforcement improves classifier system learning rate in two-class problems with unequal base rates. In: Koza J.R., Banzhaf W., Chellapilla K., et al (eds.): Genetic Programming 1998: Proceedings of the Third Annual Conference, Morgan Kaufmann, San Francisco (1998) 635–644.
Holmes J.H.: Quantitative methods for evaluating learning classifier system performance In forced two-choice decision tasks. In: Wu, A. (ed.) Proceedings of the Second International Workshop on Learning Classifier Systems (IWLCS99). Morgan Kaufmann, SanFrancisco (1999) 250–257.
Iba, H.: Bagging, Boosting, and bloating in genetic programming. In: Banzhaf, J., Daida, J., Eiben, et al (eds.): GECCO-99. Proceedings of the Genetic and Evolutionary Computation Conference. Morgan Kaufmann, San Francisco (1999) 1053–1060.
Quinlan, J.R.: C4.5: Programs for Machine Learning. Morgan Kaufmann, San Mateo, CA (1993).
Robertson, G.G. and Riolo, R.L.: A tale of two classifier systems, Machine Learning 3 (1988) 139–159.
Schapire, R.E.: Theoretical views of boosting. In: Computational Learning Theory, 4th European Conference, EuroCOLT99. Springer-Verlag, Berlin (1999) 1–10.
Ting, K.M. and Zheng, Z.: Improving the performance of boosting for naive Bayesian classification. In: Zhong, N. and Zhou, L. (eds.): Proceedings of PAKDD-00, Third Pacific-Asia Conference on Knowledge Discovery and Data Mining. Springer-Verlag, Berlin (1999) 296–305.
Weiss, S.M. and Indurkhya, N.: Predictive Data Mining. Morgan Kaufmann Publishers, Inc., San Francisco (1998).
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2000 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Holmes, J.H., Durbin, D.R., Winston, F.K. (2000). A New Bootstrapping Method to Improve Classification Performance in Learning Classifier Systems. In: Schoenauer, M., et al. Parallel Problem Solving from Nature PPSN VI. PPSN 2000. Lecture Notes in Computer Science, vol 1917. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-45356-3_73
Download citation
DOI: https://doi.org/10.1007/3-540-45356-3_73
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-41056-0
Online ISBN: 978-3-540-45356-7
eBook Packages: Springer Book Archive