Abstract
Boosting is one of the recent major developments in classification methods. The technique works by creating different versions of a classifier using an adaptive resampling procedure and then combining these classifiers using weighted voting. In this paper, several modifications of the original version of boosting, the AdaBoost algorithm introduced by Y. Freund and R.E. Schapire in 1996, will be explained. These will be shown to substantially improve the predictive power of the original version. In the first modification, weighted error estimation in AdaBoost is replaced by unweighted error estimation and this is designed to reduce the impact of observations that possess large weight. In the second modification, only a selection of base classifiers, i.e. those that contribute significantly to predictive power of the boosting model, will be included in the final model. In addition to these two modifications, we will also utilise different classification techniques as base classifiers in order to product a final boosting model. Applying these proposed modifications to three data sets from the banking industry provides results which indicate a significant and substantial improvement in predictive power over the original AdaBoost algorithm.
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Schapire, R. E.: The strength of weak learnability. Machine Learning 5 (1990) 197–227
Freund, Y., Schapire, R.: A decision-theoretic generalization of on-line learning and an application to boosting. J. Comput. System Sci. 55 (1997) 119–139
Freund, Y., Schapire, R.: Experiment with a new boosting algorithm. In Machine Learning: Proceedings of the Thirteenth International Conferences (L. Saitta, ed.). Morgan Kaufmann, San Francisco (1996) pp 148–156
Friedman, J., Hastie, T., Tibshirani, R.: Additive Logistic Regression: A Statistical View of Boosting. The Annals of Statistics 28(2) (2000) 337–407
Breiman, L.: Arcing Classifiers. The Annals of Statistics 26(3) (1998) 801–849
Krogh, A., Vedelsby, J.: Neural network ensembles, cross validation, and active learning. Tesauro, G., Touretzky, D., Leen, T. (Eds.), Advances in Neural Information Processing Systems, Vol. 7. MIT Press, Boston, MA (1995)
Webb, G. I.: MultiBoosting: A Technique for Combining Boosting and Wagging. Machine Learning. 40 (2000) 159–196
Breiman, L., Friedman, J. H., Olshen, R. A., Stone, C. J.: Classification and Regression Trees. Wadsworth Internation Group, Belmont, CA (1984)
Thomas, L. C.: A survey of credit and behavioural scoring: forecasting finance risk of lending to consumers. International Journal of Forecasting 16 (2000) 149–172
Schapire, R., Freund, Y., Bartlett, P., Lee, W. S.: Boosting the margin: A new explanation for effectiveness of voting methods. The Annals of Statistics. 26 (1998) 1651–1686
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2003 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Boonyanunta, N., Zeephongsekul, P. (2003). Improving the Predictive Power of AdaBoost: A Case Study in Classifying Borrowers. In: Chung, P.W.H., Hinde, C., Ali, M. (eds) Developments in Applied Artificial Intelligence. IEA/AIE 2003. Lecture Notes in Computer Science(), vol 2718. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-45034-3_68
Download citation
DOI: https://doi.org/10.1007/3-540-45034-3_68
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-40455-2
Online ISBN: 978-3-540-45034-4
eBook Packages: Springer Book Archive