Definition
Boosting is a kind of ensemble method which produces a strong learner that is capable of making very accurate predictions by combining rough and moderately inaccurate learners (which are called as base learners or weak learners). In particular, Boosting sequentially trains a series of base learners by using a base learning algorithm, where the training examples wrongly predicted by a base learner will receive more attention from the successive base learner. After that, it generates a final strong learner through a weighted combination of these base learners.
Historical Background
In 1988, Kearns and Valiant posed an interesting question for the research of computational learning theory, i.e., whether a weak learning algorithm that performs just slightly better than random guess can be “boosted” into an arbitrarily accurate strong learning algorithm. In other words, whether two complexity classes, weakly learnable and strongly learnableproblems, are equal. In 1989, Schapire...
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Recommended Reading
Bauer E. and Kohavi R. An empirical comparison of voting classification algorithms: Bagging, Boosting, and variants. Mach. Learn., 36(1–2):105–139, 1999.
Breiman L. Prediction games and arcing classifiers. Neural Comput., 11(7):1493–1517, 1999.
Freund Y. and Schapire R.E. Game theory, on-line prediction and Boosting. In Proc. Ninth Annual Conf. on Computational Learning Theory, 1996, pp. 325–332.
Freund Y. and Schapire R.E. A decision-theoretic generalization of on-line learning and an application to Boosting. J. Comput. Syst. Sci., 55(1):119–139, 1997. (A short version appeared in the Proceedings of EuroCOLT’95).
Friedman J., Hastie T., and Tibshirani R. Additive logistic regression: A statistical view of Boosting with discussions. Ann. Stat., 28(2):337–407, 2000.
Meir R. and Rätsch G. An introduction to Boosting and leveraging. In Advanced Lectures in Machine Learning, S. Mendelson and A.J. Smola (eds.). LNCS, Vol. 2600, Springer, Berlin, 2003, pp. 118–183.
Opitz D. and Maclin R. Popular ensemble methods: An empirical study. J. Artif. Intell. Res., 11:169–198, 1999.
Reyzin L. and Schapire R.E. How boosting the margin can also boost classifier complexity. In Proc. 23rd Int. Conf. on Machine Learning, pp. 753–760.2006,
Schapire R.E. The strength of weak learnability. Mach. Learn., 5(2):197–227, 1990.
Schapire R.E. The Boosting approach to machine learning: An overview. In Nonlinear Estimation and Classification. D.D. Denison, M.H. Hansen, C. Holmes, B. Mallick, and B. Yu (eds.). Springer, Berlin, 2003.
Schapire R.E., Freund Y., Bartlett P., and Lee W.S. Boosting the margin: A new explanation for the effectiveness of voting methods. Ann. Stat., 26(5):1651–1686, 1998.
Viola P. and Jones M. Rapid object detection using a boosted cascade of simple features. In Proc. IEEE Comp. Soc. Conf. on Computer Vision and Pattern Recognition, 2001, pp. 511–518.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2009 Springer Science+Business Media, LLC
About this entry
Cite this entry
Zhou, ZH. (2009). Boosting. In: LIU, L., ÖZSU, M.T. (eds) Encyclopedia of Database Systems. Springer, Boston, MA. https://doi.org/10.1007/978-0-387-39940-9_568
Download citation
DOI: https://doi.org/10.1007/978-0-387-39940-9_568
Publisher Name: Springer, Boston, MA
Print ISBN: 978-0-387-35544-3
Online ISBN: 978-0-387-39940-9
eBook Packages: Computer ScienceReference Module Computer Science and Engineering