Abstract
Combining the results of a number of individually trained classification systems to obtain a more accurate classifier is a widely used technique in pattern recognition. In this article, we have introduced a rough set based meta classifier (RSM). Theoretical analysis of the proposed RSM is carried out in relation to Bayes classifier since Bayes classifier is the best classifier. It has been shown that the performance of the meta classifier is at least as good as the best constituent classifier, and if one of the base classifiers of RSM converges to Bayes then RSM converges to Bayes classifier. Experimental studies show that the meta classifier improves accuracy of classification and beats other ensemble approaches in accuracy by a decisive margin, thus demonstrating the theoretical results.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Dietterich, T.G.: Machine-learning research: Four current directions. The AI Magazine 18(4), 97–136 (1998)
Pawlak, Z.: Rough Sets: Theoretical Aspects of Reasoning About Data. Kluwer Academic Publishers, Dordrecht (1991)
Wang, G., Liu, Q., Yao, Y., Skowron, A. (eds.): RSFDGrC 2003. LNCS, vol. 2639, pp. 41–48. Springer, Heidelberg (2003)
Saha, S., Murthy, C.A., Pal, S.K.: Rough set based ensemble classifier for web page classification. Fundamentae Informetica 76(1-2), 171–187 (2007)
Fukunaga, K.: Introduction to statistical pattern recognition, 2nd edn. Academic Press Professional, Inc., San Diego (1990)
Ross Quinlan, J.: C4.5: Programs for machine learning, p. 302. Morgan Kaufmann Publishers Inc., San Francisco (1993)
Clark, P., Boswell, R.: Rule induction with CN2: Some recent improvements. In: Proc. Fifth European Working Session on Learning, pp. 151–163 (1991)
Wettschereck, D., Dietterich, T.G.: An experimental comparison of the nearest-neighbor and nearest-hyperrectangle algorithms. Machine Learning 19(1), 5–27 (1995)
Cortes, C., Vapnik, V.: Support-vector networks. Machine Learning 20(3), 273–297 (1995)
Gama, J.: Iterative bayes. Theoretical Computer Science 292, 417–430 (2003)
Wolpert, D.H.: Stacked generalization. Neural Networks 5, 241–259 (1992)
Merz, C.J.: Using Correspondence Analysis to Combine Classiers. Machine Learning 36(1/2), 33–58 (1999)
Zenko, B., Todorovski, L., Dzeroski, S.: A comparison of stacking with meta decision trees to bagging, boosting, and stacking with other methods. In: ICDM 2001: Proceedings of the 2001 IEEE International Conference on Data Mining, pp. 669–670. IEEE Computer Society, Washington (2001)
Melville, P., Mooney, R.: Constructing diverse classifier ensembles using artificial training examples. In: Proc. of 18th Intl. Joint Conf. on Artificial Intelligence, Acapulco, Mexico, pp. 505–510 (2003)
Freund, Y., Schapire, R.E.: Experiments with a new boosting algorithm. In: International Conference on Machine Learning, pp. 148–156 (1996)
Breiman, L.: Bagging predictors. Machine Learning 24, 123–140 (1996)
Kuncheva, L.I., Whitaker, C.J.: Measures of diversity in classifier ensembles. Machine Learning 51, 181–207 (2003)
Best, D.J., Rayner, J.C.W.: Welch’s approximate solution for the Behrens-Fisher problem. Technometrics 29(2), 205–210 (1987)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2009 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Saha, S., Murthy, C.A., Pal, S.K. (2009). Rough Ensemble Classifier: A Comparative Study. In: Di Gesù, V., Pal, S.K., Petrosino, A. (eds) Fuzzy Logic and Applications. WILF 2009. Lecture Notes in Computer Science(), vol 5571. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-02282-1_15
Download citation
DOI: https://doi.org/10.1007/978-3-642-02282-1_15
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-02281-4
Online ISBN: 978-3-642-02282-1
eBook Packages: Computer ScienceComputer Science (R0)