Abstract
Boosting and Bagging, as two representative approaches to learning classifier committees, have demonstrated great success, especially for decision tree learning. They repeatedly build different classifiers using a base learning algorithm, by changing the distribution of the training set. Sasc, as a different type of committee learning method, can also significantly reduce the error rate of decision trees. It generates classifier committees by stochastically modifying the set of attributes but keeping the distribution of the training set unchanged. It has been shown that Bagging and Sasc are, on average, less accurate than Boosting, but the performance of the former is more stable than that of the latter in terms of less frequently obtaining significantly higher error rates than the base learning algorithm. In this paper, we propose a novel committee learning algorithm, called SascBag, that combines Sasc and Bagging. It creates different classifiers by stochastically varying both the attribute set and the distribution of the training set. Experimental results in a representative collection of natural domains show that, for decision tree learning, the new algorithm is, on average, more accurate than Boosting, Bagging, and Sasc. It is more stable than Boosting. In addition, like Bagging and Sasc, SascBag is amenable to parallel and distributed processing while Boosting is not. This gives SascBag another advantage over Boosting for parallel machine learning and datamining.
Preview
Unable to display preview. Download preview PDF.
References
Freund, Y.: Boosting a weak learning algorithm by majority. Information and Computation 121 (1996) 256–285.
Freund, Y. and Schapire, R.E.: A decision-theoretic generalization of on-line learning and an application to Boosting. Unpublished manuscript (1996a) (available at http://www.research.att.com/~yoav).
Freund, Y. and Schapire, R.E.: Experiments with a new Boosting algorithm. Proceedings of the Thirteenth International Conference on Machine Learning. San Francisco, CA: Morgan Kaufmann (1996b) 148–156.
Quinlan, J.R.: Bagging Boosting, and C4.5: Proceedings of the Thirteenth National Conference on Artificial Intelligence. Menlo Park, CA: AAAI Press (1996) 725–730.
Breiman, L.: Bagging predictors. Machine Learning 24 (1996a) 123–140.
Breiman, L.: Arcing classifiers. Technical Report (available at: http://www.stat. Berkeley.EDU/users/breiman/) Department of Statistics, University of California, Berkeley, CA (1996b).
Dietterich, T.G. and Kong, E.B.: Machine learning bias, statistical bias, and statistical variance of decision tree algorithms. Technical Report, Dept of Computer Science, Oregon State University, Corvallis, Oregon (1995) (available at ftp://ftp.cs.orst.edu/pub/tgd/papers/tr-bias.ps.gz).
Ali, K.M.: Learning Probabilistic Relational Concept Descriptions. PhD. Thesis, Dept of Info. and Computer Science, Univ. of California, Irvine (1996).
Chan, P., Stolfo, S., and Wolpert, D. (eds): Working Notes of AAAI Workshop on Integrating Multiple Learned Models for Improving and Scaling Machine Learning Algorithms (available at http://www.cs.fit.edu/~imlm/papers.html), Portland, Oregon (1996).
Ali, K.M. and Pazzani, M.J.: Error reduction through learning multiple descriptions. Machine Learning 24 (1996) 173–202.
Schapire, R.E., Freund, Y., Bartlett, P., and Lee, W.S.: Boosting the margin: A new explanation for the effectiveness of voting methods. Proceedings of the Fourteenth International Conference on Machine Learning. Morgan Kaufmann (1997) 322–330.
Bauer, E. and Kohavi, R.: An empirical comparison of voting classification algorithms: Bagging, Boosting, and variants. To appear in Machine Learning (1998) (available at: http://reality.sgi.com/ronnyk/vote.ps.gz).
Schapire, R.E.: The strength of weak learnability. Machine Learning 5 (1990) 197–227.
Kwok, S.W. and Carter, C.: Multiple decision trees. Schachter, R.D., Levitt, T.S., Kanal, L.N., and Lemmer, J.F. (eds) Uncertainty in Artificial intelligence. Elsevier Science (1990) 327–335.
Dietterich, T.G. and Bakiri, G.: Solving multiclass learning problems via error-correcting output codes. Journal of Artificial Intelligence Research 2, 1995) 263–286.
Zheng, Z. and Webb, G.I.: Stochastic attribute selection committees. Technical Report (TR C98/08), School of Computing and Mathematics, Deakin University, Australia (1998) (available at http://www3.cm.deakin.edu.au/~zijian/Papers/sasc-tr-C98-08.ps.gz).
Buntine, W.: A Theory of Learning Classification Rules. PhD. Thesis, School of Computing Science, University of Technology, Sydney (1990).
Kohavi, R. and Kunz, C.: Option decision trees with majority votes. Proceedings of the Fourteenth International Conference on Machine Learning. San Francisco, CA: Morgan Kaufmann (1997) 161–169.
Cherkauer, K.J.: Human expert-level performance on a science image analysis task by a system using combined artificial neural networks. Chan, P., Stolfo, S., and Wolpert, D. (eds) Working Notes of AAAI Workshop on Integrating Multiple Learned Models for Improving and Scaling Machine Learning Algorithms (available at http://www.cs.fit.edu/~imlm/papers.html), Portland, Oregon (1996) 15–21.
Tumer, K. and Ghosh, J.: Error correction and error reduction in ensemble classifiers. Connection Science 8 (1996) 385–404.
Zheng, Z.: Naive Bayesian classifier committees. Proceedings of the Tenth European Conference on Machine Learning. Berlin: Springet-Verlag (1998) 196–207.
Asker, L. and Maclin, R.: Ensembles as a sequence of classifiers. Proccedings, of the Fifteenth International Joint Conference on Artificial Intelligence. San Francisco, CA: Morgan Kaufmann (1997) 860–865.
Wolpert, D.H.: Stacked generalization. Neural Networks 5 (1992) 241–259.
Zhang, X., Mesirrov, J.P. and Waltz, D.L.: Hybrid system for protein secondary structure prediction. Journal of Molecular Biology 225 (1992) 1049–1063.
Dietterich, T.G.: Machine learning research. AI Magazine 18 (1997) 97–136.
Quinlan, J.R.: C4.5: Program for Machine Learning. San Mateo, CA: Morgan Kaufmann (1993).
Merz, C.J. and Murphy, P.M.: UCI Repository of machine learning databases [http://www.ics.uci.edu/~mlearn/MLRepository.html]. Irvine, CA: Univ of California, Dept. of Info and Computer Science (1997).
Kohavi, R.: A study of cross-validation and bootstrap for accuracy estimation and model selection. Proceedings of the Fourteenth International Joint Conference on Artificial Intelligence. San Mateo, CA: Morgan Kaufmann (1995) 1137–1143.
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 1998 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Zheng, Z. (1998). Generating classifier committees by stochastically selecting both attributes and training examples. In: Lee, HY., Motoda, H. (eds) PRICAI’98: Topics in Artificial Intelligence. PRICAI 1998. Lecture Notes in Computer Science, vol 1531. Springer, Berlin, Heidelberg . https://doi.org/10.1007/BFb0095254
Download citation
DOI: https://doi.org/10.1007/BFb0095254
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-65271-7
Online ISBN: 978-3-540-49461-4
eBook Packages: Springer Book Archive