Skip to main content

Learning SVMs from Sloppily Labeled Data

  • Conference paper
Artificial Neural Networks – ICANN 2009 (ICANN 2009)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 5768))

Included in the following conference series:

Abstract

This paper proposes a modelling of Support Vector Machine (SVM) learning to address the problem of learning with sloppy labels. In binary classification, learning with sloppy labels is the situation where a learner is provided with labelled data, where the observed labels of each class are possibly noisy (flipped) version of their true class and where the probability of flipping a label y to –y only depends on y. The noise probability is therefore constant and uniform within each class: learning with positive and unlabeled data is for instance a motivating example for this model. In order to learn with sloppy labels, we propose SloppySvm, an SVM algorithm that minimizes a tailored nonconvex functional that is shown to be a uniform estimate of the noise-free SVM functional. Several experiments validate the soundness of our approach.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Bartlett, P.L., Mendelson, S.: Rademacher and gaussian complexities: Risk bounds and structural results. J. of Machine Learning Research 3, 463–482 (2002)

    MathSciNet  MATH  Google Scholar 

  2. Blum, A., Mitchell, T.: Combining Labeled and Unlabeled Data with Co-Training. In: Proc. of the 11th Conf. on Computational Learning Theory, pp. 92–100 (1998)

    Google Scholar 

  3. Chapelle, O.: Training a support vector machine in the primal. Neural Comput. 19(5), 1155–1178 (2007)

    Article  MathSciNet  MATH  Google Scholar 

  4. Magnan, C.: Asymmetrical Semi-Supervised Learning and Prediction of Disulfide Connectivity in Proteins. RIA, New Methods in Machine Learning: Theory and applications 20(6), 673–695 (2006)

    Google Scholar 

  5. McDiarmid, C.: On the method of bounded differences. In: Survey in Combinatorics, pp. 148–188 (1989)

    Google Scholar 

  6. Schölkopf, B., Smola, A.J.: Learning with Kernels, Support Vector Machines, Regularization, Optimization and Beyond. MIT University Press, Cambridge (2002)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2009 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Stempfel, G., Ralaivola, L. (2009). Learning SVMs from Sloppily Labeled Data. In: Alippi, C., Polycarpou, M., Panayiotou, C., Ellinas, G. (eds) Artificial Neural Networks – ICANN 2009. ICANN 2009. Lecture Notes in Computer Science, vol 5768. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-04274-4_91

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-04274-4_91

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-04273-7

  • Online ISBN: 978-3-642-04274-4

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics