Skip to main content

Stochastic Attribute Selection Committees

  • Scientific Track
  • Conference paper
  • First Online:
Advanced Topics in Artificial Intelligence (AI 1998)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 1502))

Included in the following conference series:

Abstract

Classifier committee learning methods generate multiple classifiers to form a committee by repeated application of a single base learning algorithm. The committee members vote to decide the final classification. Two such methods, Bagging and Boosting, have shown great success with decision tree learning. They create different classifiers by modifying the distribution of the training set. This paper studies a different approach: Stochastic Attribute Selection Committee learning of decision trees. It generates classifier committees by stochastically modifying the set of attributes but keeping the distribution of the training set unchanged. An empirical evaluation of a variant of this method, namely Sasc, in a representative collection of natural domains shows that the SASC method can significantly reduce the error rate of decision tree learning. On average Sasc is more accurate than Bagging and less accurate than Boosting, although a one-tailed sign-test fails to show that these differences are significant at a level of 0.05. In addition, it is found that, like Bagging, Sasc is more stable than Boosting in terms of less frequently obtaining significantly higher error rates than C4.5 and, when error is raised, producing lower error rate increases. Moreover, like Bagging, Sasc is amenable to parallel and distributed processing while Boosting is not.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  • Ali, K.M.: Learning Probabilistic Relational Concept Descriptions. PhD. Thesis, Dept of Info. and Computer Science, Univ. of California, Irvine (1996).

    Google Scholar 

  • Ali, K.M. and Pazzani, M.J.: Error reduction through learning multiple descriptions. Machine Learning 24 (1996) 173–202.

    Google Scholar 

  • Bauer, E. and Kohavi, R.: An empirical comparison of voting classification algorithms: Bagging, Boosting, and variants. Submitted to Machine Learning (1998) (available at: http://reality.sgi.com/ronnyk/vote.ps.gz).

    Google Scholar 

  • Breiman, L.: Bagging predictors. Machine Learning 24 (1996a) 123–140.

    MATH  MathSciNet  Google Scholar 

  • Breiman, L.: Arcing classifiers. Technical Report (available at: http://www.stat. Berkeley.EDU/users/breiman/). Department of Statistics, University of California, Berkeley, CA (1996b).

    Google Scholar 

  • Buntine, W.: A Theory of Learning Classification Rules. PhD. Thesis, School of Computing Science, University of Technology, Sydney (1990).

    Google Scholar 

  • Chan, P., Stolfo, S., and Wolpert, D. (eds): Working Notes of AAAI Workshop on Integrating Multiple Learned Models for Improving and Scaling Machine Learning Algorithms (available at http://www.cs.fit.edu/~imlm/papers.html), Portland, Oregon (1996).

    Google Scholar 

  • Cherkauer, K.J.: Human expert-level performance on a science image analysis task by a system using combined artificial neural networks. Chan, P., Stolfo, S., and Wolpert, D. (eds) Working Notes of AAAI Workshop on Integrating Multiple Learned Models for Improving and Scaling Machine Learning Algorithms (available at http://www.cs.fit.edu/~imlm/papers.html), Portland, Oregon (1996) 15–21.

    Google Scholar 

  • Dietterich, T.G. and Bakiri, G.: Solving multiclass learning problems via error-correcting output codes. Journal of AI Research 2 (1995) 263–286.

    MATH  Google Scholar 

  • Dietterich, T.G. and Kong, E.B.: Machine learning bias, statistical bias, and statistical variance of decision tree algorithms. Technical Report, Dept of Computer Science, Oregon State University, Corvallis, Oregon (1995) (available at ftp://ftp.cs.orst.edu/pub/tgd/papers/tr-bias.ps.gz)

    Google Scholar 

  • Dietterich, T. G.: Machine learning research. AI Magazine 18 (1997) 97–136.

    Google Scholar 

  • Freund, Y.: Boosting a weak learning algorithm by majority. Information and Computation 121 (1996) 256–285.

    Article  MathSciNet  Google Scholar 

  • Freund, Y. and Schapire, R.E.: Experiments with a new Boosting algorithm. Proceedings of the Thirteenth International Conference on Machine Learning. San Francisco, CA: Morgan Kaufmann (1996) 148–156.

    Google Scholar 

  • Kohavi, R. and Kunz, C.: Option decision trees with majority votes. Proceedings of the Fourteenth International Conference on Machine Learning. San Francisco, CA: Morgan Kaufmann (1997) 161–169.

    Google Scholar 

  • Kwok, S.W. and Carter, C.: Multiple decision trees. Schachter, R.D., Levitt, T.S., Kanal, L.N., and Lemmer, J.F. (eds) Uncertainty in Artificial Intelligence. Elsevier Science (1990) 327–335.

    Google Scholar 

  • Merz, C.J. and Murphy, P.M.: UCI Repository of machine learning databases [http://www.ics.uci.edu/~mlearn/MLRepository.html]. Irvine, CA: Univ of California, Dept of Info and Computer Science (1997).

    Google Scholar 

  • Quinlan, J.R.: C4.5: Program for Machine Learning. Morgan Kaufmann (1993).

    Google Scholar 

  • Quinlan, J.R.: Bagging, Boosting, and C4.5. Proceedings of the 13th National Conference on Artificial Intelligence. Menlo Park, CA: AAAI Press (1996) 725–730.

    Google Scholar 

  • Schapire, R.E.: The strength of weak learnability. Machine Learning 5 (1990) 197–227.

    Google Scholar 

  • Schapire, R.E., Freund, Y., Bartlett, P., and Lee, W.S.: Boosting the margin: A new explanation for the effectiveness of voting methods. Proceedings of the Fourteenth International Conference on Machine Learning. Morgan Kaufmann (1997) 322–330.

    Google Scholar 

  • Tumer, K. and Ghosh, J.: Error correction and error reduction in ensemble classifiers. Connection Science 8 (1996) 385–404.

    Article  Google Scholar 

  • Wolpert, D.H.: Stacked generalization. Neural Networks 5 (1992) 241–259.

    Article  Google Scholar 

  • Zheng, Z.: Naive Bayesian classifier committees. Proceedings of the 10th European Conference on Machine Learning. Berlin: Springet-Verlag (1998) 196–207.

    Google Scholar 

  • Zheng, Z. and Webb, G.I.: Stochastic attribute selection committees. Technical Report (TR C98/08), School of Computing and Mathematics, Deakin University, Australia (1998) (available at http://www3.cm.deakin.edu.au/~zijian/Papers/sasc-tr-C98-08.ps.gz)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Grigoris Antoniou John Slaney

Rights and permissions

Reprints and permissions

Copyright information

© 1998 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Zheng, Z., Webb, G.I. (1998). Stochastic Attribute Selection Committees. In: Antoniou, G., Slaney, J. (eds) Advanced Topics in Artificial Intelligence. AI 1998. Lecture Notes in Computer Science, vol 1502. Springer, Berlin, Heidelberg . https://doi.org/10.1007/BFb0095063

Download citation

  • DOI: https://doi.org/10.1007/BFb0095063

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-65138-3

  • Online ISBN: 978-3-540-49561-1

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics