Abstract.
In this paper we study methods that combine multiple classification models learned over separate data sets. Numerous studies posit that such approaches provide the means to efficiently scale learning to large data sets, while also boosting the accuracy of individual classifiers. These gains, however, come at the expense of an increased demand for run-time system resources. The final ensemble meta-classifier may consist of a large collection of base classifiers that require increased memory resources while also slowing down classification throughput. Here, we describe an algorithm for pruning (i.e., discarding a subset of the available base classifiers) the ensemble meta-classifier as a means to reduce its size while preserving its accuracy and we present a technique for measuring the trade-off between predictive performance and available run-time system resources. The algorithm is independent of the method used initially when computing the meta-classifier. It is based on decision tree pruning methods and relies on the mapping of an arbitrary ensemble meta-classifier to a decision tree model. Through an extensive empirical study on meta-classifiers computed over two real data sets, we illustrate our pruning algorithm to be a robust and competitive approach to discarding classification models without degrading the overall predictive performance of the smaller ensemble computed over those that remain after pruning.
Similar content being viewed by others
Author information
Authors and Affiliations
Additional information
Received 30 August 2000 / Revised 7 March 2001 / Accepted in revised form 21 May 2001
Rights and permissions
About this article
Cite this article
Prodromidis, A., Stolfo, S. Cost Complexity-Based Pruning of Ensemble Classifiers. Knowledge and Information Systems 3, 449–469 (2001). https://doi.org/10.1007/PL00011678
Issue Date:
DOI: https://doi.org/10.1007/PL00011678