Abstract
Since their invention by Vladimir Vapnik and his co-workers in the early 1990s, support vector machines (SVMs)Support vector machine (SVM)—( have attracted a lot of research activities from various communities. While at the beginning this research mostly focused on generalization bounds, the last decade witnessed a shift towards consistencyConsistency, oracle inequalities, and learning ratesLearning rate. We discuss some of these developments in view of binary classificationBinary classification and least squares regressionLeast squares regression.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Aizerman, M., Braverman, E., Rozonoer, L.: Theoretical foundations of the potential function method in pattern recognition learning. Autom. Remote Control 25, 821–837 (1964)
Bartlett, P., Jordan, M., McAuliffe, J.: Convexity, classification, and risk bounds. J. Am. Stat. Assoc. 101, 138–156 (2006)
Bauer, H.: Measure and Integration Theory. De Gruyter, Berlin (2001)
Bennett, C., Sharpley, R.: Interpolation of Operators. Academic, Boston (1988)
Bergh, J., Löfström, J.: Interpolation Spaces: An Introduction. Springer, New York (1976)
Boser, B., Guyon, I., Vapnik, V.: A training algorithm for optimal margin classifiers. In: Proceedings of the 5th Annual ACM Workshop on Computational Learning Theory, Pittsburgh, pp. 144–152 (1992)
Caponnetto, A., De Vito, E.: Optimal rates for regularized least squares algorithm. Found. Comput. Math. 7, 331–368 (2007)
Cortes, C., Vapnik, V.: Support vector networks. Mach. Learn. 20, 273–297 (1995)
Cucker, F., Smale, S.: On the mathematical foundations of learning. Bull. Am. Math. Soc. 39, 1–49 (2002)
Devroye, L.: Any discrimination rule can have an arbitrarily bad probability of error for finite sample size. IEEE Trans. Pattern Anal. Mach. Intell. 4, 154–157 (1982)
Devroye, L., Györfi, L., Lugosi, G.: A Probabilistic Theory of Pattern Recognition. Springer, New York (1996)
Drucker, H., Burges, C., Kaufman, L., Smola, A., Vapnik, V.: Support vector regression machines. In: Advances in Neural Information Processing Systems, Denver, vol. 9, pp. 155–161. MIT, Cambridge (1997)
Eberts, M., Steinwart, I.: Optimal regression rates for SVMs using Gaussian kernels. Electron. J. Stat. 7, 1–42 (2013)
Edmunds, D., Triebel, H.: Function Spaces, Entropy Numbers, Differential Operators. Cambridge University Press, Cambridge (1996)
Györfi, L., Kohler, M., Krzyżak, A., Walk, H.: A Distribution-Free Theory of Nonparametric Regression. Springer, New York (2002)
Kimeldorf, G., Wahba, G.: Some results on Tchebycheffian spline functions. J. Math. Anal. Appl. 33, 82–95 (1971)
Kohler, M., Krzyżak, A.: Nonparametric regression estimation using penalized least squares. IEEE Trans. Inform. Theory 47, 3054–3058 (2001)
Kohler, M., Krzyżak, A., Schäfer, D.: Application of structural risk minimization to multivariate smoothing spline regression estimates. Bernoulli 4, 475–489 (2002)
Koltchinskii, V., Beznosova, O.: Exponential convergence rates in classification. In: Proceedings of the 18th Annual Conference on Learning Theory, Bertinoro, pp. 295–307 (2005)
Loustau, S.: Aggregation of SVM classifiers using Sobolev spaces. J. Mach. Learn. Res. 9, 1559–1582 (2008)
Mendelson, S., Neeman, J.: Regularization in kernel learning. Ann. Stat. 38, 526–565 (2010)
Micchelli, C., Xu, Y., Zhang, H.: Universal kernels. J. Mach. Learn. Res. 7, 2651–2667 (2006)
Poggio, T., Girosi, F.: A theory of networks for approximation and learning. Proc. IEEE 78, 1481–1497 (1990)
Silverman, B.: Some aspects of the spline smoothing approach to nonparametric regression. J. Royal Stat. Soc. B Stat. Methodol. 47, 1–52 (1985)
Smale, S., Zhou, D.: Estimating the approximation error in learning theory. Anal. Appl. 1, 17–41 (2003)
Steinwart, I.: On the influence of the kernel on the consistency of support vector machines. J. Mach. Learn. Res. 2, 67–93 (2001)
Steinwart, I.: Support vector machines are universally consistent. J. Complexity 18, 768–791 (2002)
Steinwart, I., Christmann, A.: Support Vector Machines. Springer, New York (2008)
Steinwart, I., Christmann, A.: Estimating conditional quantiles with the help of the pinball loss. Bernoulli 17, 211–225 (2011)
Steinwart, I., Scovel, C.: Fast rates for support vector machines using Gaussian kernels. Ann. Stat. 35, 575–607 (2007)
Steinwart, I., Scovel, C.: Mercer’s theorem on general domains: on the interaction between measures, kernels, and RKHSs. Constr. Approx. 35, 363–417 (2012)
Steinwart, I., Hush, D., Scovel, C.: Optimal rates for regularized least squares regression. In: Proceedings of the 22nd Annual Conference on Learning Theory, Montreal, pp. 79–93 (2009)
Stone, C.: Consistent nonparametric regression. Ann. Stat. 5, 595–645 (1977)
Tsybakov, A.: Optimal aggregation of classifiers in statistical learning. Ann. Stat. 32, 135–166 (2004)
van de Geer, S.: A new approach to least squares estimation, with applications. Ann. Stat. 15, 587–602 (1987)
Vapnik, V.: The Nature of Statistical Learning Theory. Springer, New York (1995)
Vapnik, V., Lerner, A.: Pattern recognition using generalized portrait method. Autom. Remote Control 24, 774–780 (1963)
Vapnik, V., Golowich, S., Smola, A.: Support vector method for function approximation, regression estimation, and signal processing. In: Advances in Neural Information Processing Systems, Denver, vol. 9, pp. 81–287. MIT, Cambridge (1997)
Wahba, G.: Spline Models for Observational Data. Series in Applied Mathematics, vol. 59. SIAM, Philadelphia (1990)
Zhang, T.: Statistical behaviour and consistency of classification methods based on convex risk minimization. Ann. Stat. 32, 56–134 (2004)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer-Verlag Berlin Heidelberg
About this chapter
Cite this chapter
Steinwart, I. (2013). Some Remarks on the Statistical Analysis of SVMs and Related Methods. In: Schölkopf, B., Luo, Z., Vovk, V. (eds) Empirical Inference. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-41136-6_4
Download citation
DOI: https://doi.org/10.1007/978-3-642-41136-6_4
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-41135-9
Online ISBN: 978-3-642-41136-6
eBook Packages: Computer ScienceComputer Science (R0)