Skip to main content

Normalization in Support Vector Machines

  • Conference paper
  • First Online:
Pattern Recognition (DAGM 2001)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 2191))

Included in the following conference series:

Abstract

This article deals with various aspects of normalization in the context of Support Vector Machines. We consider fist normalization of the vectors in the input space and point out the inherent limitations. A natural extension to the feature space is then represented by the kernel function normalization. A correction of the position of the Optimal Separating Hyperplane is subsequently introduced so as to suit better these normalized kernels. Numerical experiments finally evaluate the different approaches.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. B. Schölkopf. Support Vector Learning. R. Oldenburg Verlag, Munich, 1997.

    MATH  Google Scholar 

  2. B. Schölkopf, C.J.C. Burges & A.J. Smola. Advances in Kernel Methods. The MIT Press, MA, 1999.

    Google Scholar 

  3. C. Cortes & V. Vapnik. Support-Vector Networks. Machine Learning, Kluwer Academic Publishers, 273–297, 1995.

    Google Scholar 

  4. S. Haykins. Neural Networks: a Comprehensive Approach. Prentice Hall, 1999.

    Google Scholar 

  5. R. Herbrich & T. Graepel. A PAC-Bayesian Margin Bound for Linear Classifiers: WhySVMs work. Advances in Neural Information System Processing 13, 2001 (in press).

    Google Scholar 

  6. B. Schölkopf, J. C. Platt, J. Shawe-Taylor, A. Smola & R. C. Williamson. Estimating the Support of a High-Dimensional Distribution. Technical Report MSR-TR-99-87, 2000.

    Google Scholar 

  7. M. Pontil & A. Verri. Support Vector Machines for 3D Object Recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence 20, 637–646, 1998.

    Article  Google Scholar 

  8. W. Maass. On the Computational Power of Winner-Take-All. Neural Computation 12, 2519–2535, 2000.

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2001 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Graf, A.B., Borer, S. (2001). Normalization in Support Vector Machines. In: Radig, B., Florczyk, S. (eds) Pattern Recognition. DAGM 2001. Lecture Notes in Computer Science, vol 2191. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-45404-7_37

Download citation

  • DOI: https://doi.org/10.1007/3-540-45404-7_37

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-42596-0

  • Online ISBN: 978-3-540-45404-5

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics