Skip to main content

Generating classifier committees by stochastically selecting both attributes and training examples

  • Induction (Improving Classifier’s Accuracy)
  • Conference paper
  • First Online:
PRICAI’98: Topics in Artificial Intelligence (PRICAI 1998)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 1531))

Included in the following conference series:

Abstract

Boosting and Bagging, as two representative approaches to learning classifier committees, have demonstrated great success, especially for decision tree learning. They repeatedly build different classifiers using a base learning algorithm, by changing the distribution of the training set. Sasc, as a different type of committee learning method, can also significantly reduce the error rate of decision trees. It generates classifier committees by stochastically modifying the set of attributes but keeping the distribution of the training set unchanged. It has been shown that Bagging and Sasc are, on average, less accurate than Boosting, but the performance of the former is more stable than that of the latter in terms of less frequently obtaining significantly higher error rates than the base learning algorithm. In this paper, we propose a novel committee learning algorithm, called SascBag, that combines Sasc and Bagging. It creates different classifiers by stochastically varying both the attribute set and the distribution of the training set. Experimental results in a representative collection of natural domains show that, for decision tree learning, the new algorithm is, on average, more accurate than Boosting, Bagging, and Sasc. It is more stable than Boosting. In addition, like Bagging and Sasc, SascBag is amenable to parallel and distributed processing while Boosting is not. This gives SascBag another advantage over Boosting for parallel machine learning and datamining.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Freund, Y.: Boosting a weak learning algorithm by majority. Information and Computation 121 (1996) 256–285.

    Article  MathSciNet  Google Scholar 

  2. Freund, Y. and Schapire, R.E.: A decision-theoretic generalization of on-line learning and an application to Boosting. Unpublished manuscript (1996a) (available at http://www.research.att.com/~yoav).

    Google Scholar 

  3. Freund, Y. and Schapire, R.E.: Experiments with a new Boosting algorithm. Proceedings of the Thirteenth International Conference on Machine Learning. San Francisco, CA: Morgan Kaufmann (1996b) 148–156.

    Google Scholar 

  4. Quinlan, J.R.: Bagging Boosting, and C4.5: Proceedings of the Thirteenth National Conference on Artificial Intelligence. Menlo Park, CA: AAAI Press (1996) 725–730.

    Google Scholar 

  5. Breiman, L.: Bagging predictors. Machine Learning 24 (1996a) 123–140.

    MATH  MathSciNet  Google Scholar 

  6. Breiman, L.: Arcing classifiers. Technical Report (available at: http://www.stat. Berkeley.EDU/users/breiman/) Department of Statistics, University of California, Berkeley, CA (1996b).

    Google Scholar 

  7. Dietterich, T.G. and Kong, E.B.: Machine learning bias, statistical bias, and statistical variance of decision tree algorithms. Technical Report, Dept of Computer Science, Oregon State University, Corvallis, Oregon (1995) (available at ftp://ftp.cs.orst.edu/pub/tgd/papers/tr-bias.ps.gz).

    Google Scholar 

  8. Ali, K.M.: Learning Probabilistic Relational Concept Descriptions. PhD. Thesis, Dept of Info. and Computer Science, Univ. of California, Irvine (1996).

    Google Scholar 

  9. Chan, P., Stolfo, S., and Wolpert, D. (eds): Working Notes of AAAI Workshop on Integrating Multiple Learned Models for Improving and Scaling Machine Learning Algorithms (available at http://www.cs.fit.edu/~imlm/papers.html), Portland, Oregon (1996).

    Google Scholar 

  10. Ali, K.M. and Pazzani, M.J.: Error reduction through learning multiple descriptions. Machine Learning 24 (1996) 173–202.

    Google Scholar 

  11. Schapire, R.E., Freund, Y., Bartlett, P., and Lee, W.S.: Boosting the margin: A new explanation for the effectiveness of voting methods. Proceedings of the Fourteenth International Conference on Machine Learning. Morgan Kaufmann (1997) 322–330.

    Google Scholar 

  12. Bauer, E. and Kohavi, R.: An empirical comparison of voting classification algorithms: Bagging, Boosting, and variants. To appear in Machine Learning (1998) (available at: http://reality.sgi.com/ronnyk/vote.ps.gz).

    Google Scholar 

  13. Schapire, R.E.: The strength of weak learnability. Machine Learning 5 (1990) 197–227.

    Google Scholar 

  14. Kwok, S.W. and Carter, C.: Multiple decision trees. Schachter, R.D., Levitt, T.S., Kanal, L.N., and Lemmer, J.F. (eds) Uncertainty in Artificial intelligence. Elsevier Science (1990) 327–335.

    Google Scholar 

  15. Dietterich, T.G. and Bakiri, G.: Solving multiclass learning problems via error-correcting output codes. Journal of Artificial Intelligence Research 2, 1995) 263–286.

    MATH  Google Scholar 

  16. Zheng, Z. and Webb, G.I.: Stochastic attribute selection committees. Technical Report (TR C98/08), School of Computing and Mathematics, Deakin University, Australia (1998) (available at http://www3.cm.deakin.edu.au/~zijian/Papers/sasc-tr-C98-08.ps.gz).

    Google Scholar 

  17. Buntine, W.: A Theory of Learning Classification Rules. PhD. Thesis, School of Computing Science, University of Technology, Sydney (1990).

    Google Scholar 

  18. Kohavi, R. and Kunz, C.: Option decision trees with majority votes. Proceedings of the Fourteenth International Conference on Machine Learning. San Francisco, CA: Morgan Kaufmann (1997) 161–169.

    Google Scholar 

  19. Cherkauer, K.J.: Human expert-level performance on a science image analysis task by a system using combined artificial neural networks. Chan, P., Stolfo, S., and Wolpert, D. (eds) Working Notes of AAAI Workshop on Integrating Multiple Learned Models for Improving and Scaling Machine Learning Algorithms (available at http://www.cs.fit.edu/~imlm/papers.html), Portland, Oregon (1996) 15–21.

    Google Scholar 

  20. Tumer, K. and Ghosh, J.: Error correction and error reduction in ensemble classifiers. Connection Science 8 (1996) 385–404.

    Article  Google Scholar 

  21. Zheng, Z.: Naive Bayesian classifier committees. Proceedings of the Tenth European Conference on Machine Learning. Berlin: Springet-Verlag (1998) 196–207.

    Google Scholar 

  22. Asker, L. and Maclin, R.: Ensembles as a sequence of classifiers. Proccedings, of the Fifteenth International Joint Conference on Artificial Intelligence. San Francisco, CA: Morgan Kaufmann (1997) 860–865.

    Google Scholar 

  23. Wolpert, D.H.: Stacked generalization. Neural Networks 5 (1992) 241–259.

    Article  Google Scholar 

  24. Zhang, X., Mesirrov, J.P. and Waltz, D.L.: Hybrid system for protein secondary structure prediction. Journal of Molecular Biology 225 (1992) 1049–1063.

    Article  Google Scholar 

  25. Dietterich, T.G.: Machine learning research. AI Magazine 18 (1997) 97–136.

    Google Scholar 

  26. Quinlan, J.R.: C4.5: Program for Machine Learning. San Mateo, CA: Morgan Kaufmann (1993).

    Google Scholar 

  27. Merz, C.J. and Murphy, P.M.: UCI Repository of machine learning databases [http://www.ics.uci.edu/~mlearn/MLRepository.html]. Irvine, CA: Univ of California, Dept. of Info and Computer Science (1997).

    Google Scholar 

  28. Kohavi, R.: A study of cross-validation and bootstrap for accuracy estimation and model selection. Proceedings of the Fourteenth International Joint Conference on Artificial Intelligence. San Mateo, CA: Morgan Kaufmann (1995) 1137–1143.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Hing-Yan Lee Hiroshi Motoda

Rights and permissions

Reprints and permissions

Copyright information

© 1998 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Zheng, Z. (1998). Generating classifier committees by stochastically selecting both attributes and training examples. In: Lee, HY., Motoda, H. (eds) PRICAI’98: Topics in Artificial Intelligence. PRICAI 1998. Lecture Notes in Computer Science, vol 1531. Springer, Berlin, Heidelberg . https://doi.org/10.1007/BFb0095254

Download citation

  • DOI: https://doi.org/10.1007/BFb0095254

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-65271-7

  • Online ISBN: 978-3-540-49461-4

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics