Skip to main content

Hinge Loss Projection for Classification

  • Conference paper
  • First Online:
Neural Information Processing (ICONIP 2016)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 9948))

Included in the following conference series:

Abstract

Hinge loss is one-sided function which gives optimal solution than that of squared error (SE) loss function in case of classification. It allows data points which have a value greater than 1 and less than \(-1\) for positive and negative classes, respectively. These have zero contribution to hinge function. However, in the most classification tasks, least square (LS) method such as ridge regression uses SE instead of hinge function. In this paper, a simple projection method is used to minimize hinge loss function through LS methods. We modify the ridge regression and its kernel based version i.e. kernel ridge regression so that it can adopt to hinge function instead of using SE in case of classification problem. The results show the effectiveness of hinge loss projection method especially on imbalanced data sets in terms of geometric mean (GM).

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    The Matlab code implementation of HLP is available in the author’s repository.

  2. 2.

    Available in http://sci2s.ugr.es/keel/imbalanced.php#sub20.

References

  1. Alcalá, J., Fernández, A., Luengo, J., Derrac, J., García, S., Sánchez, L., Herrera, F.: Keel data-mining software tool: data set repository, integration of algorithms and experimental analysis framework. J. Multiple-Valued Log. Soft Comput. 17(11), 255–287 (2011)

    Google Scholar 

  2. Barandela, R., Sánchez, J.S., Garcıa, V., Rangel, E.: Strategies for learning in class imbalance problems. Pattern Recogn. 36(3), 849–851 (2003)

    Article  Google Scholar 

  3. Friedman, J., Hastie, T., Tibshirani, R.: The Elements of Statistical Learning. Springer Series in Statistics, vol. 1. Springer, Berlin (2001)

    MATH  Google Scholar 

  4. Hoerl, A.E., Kennard, R.W.: Ridge regression: biased estimation for nonorthogonal problems. Technometrics 12(1), 55–67 (1970)

    Article  MATH  Google Scholar 

  5. Lang, K.J., Witbrock, M.J.: Learning to tell two spirals apart. In: Proceeding of 1988 Connectionist Models Summer School (1988)

    Google Scholar 

  6. Rosasco, L., De Vito, E., Caponnetto, A., Piana, M., Verri, A.: Are loss functions all the same? Neural Comput. 16(5), 1063–1076 (2004)

    Article  MATH  Google Scholar 

  7. Saunders, C., Gammerman, A., Vovk, V.: Ridge regression learning algorithm in dual variables. In: (ICML-1998) Proceedings of the 15th International Conference on Machine Learning, pp. 515–521. Morgan Kaufmann (1998)

    Google Scholar 

  8. Suykens, J.A., Vandewalle, J.: Least squares support vector machine classifiers. Neural Process. Lett. 9(3), 293–300 (1999)

    Article  MATH  Google Scholar 

  9. Vapnik, V.: The Nature of Statistical Learning Theory. Springer, New York (1995)

    Book  MATH  Google Scholar 

  10. Zhang, S., Hu, Q., Xie, Z., Mi, J.: Kernel ridge regression for general noise model with its application. Neurocomputing 149, 836–846 (2015)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Syukron Abu Ishaq Alfarozi .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2016 Springer International Publishing AG

About this paper

Cite this paper

Alfarozi, S.A.I., Woraratpanya, K., Pasupa, K., Sugimoto, M. (2016). Hinge Loss Projection for Classification. In: Hirose, A., Ozawa, S., Doya, K., Ikeda, K., Lee, M., Liu, D. (eds) Neural Information Processing. ICONIP 2016. Lecture Notes in Computer Science(), vol 9948. Springer, Cham. https://doi.org/10.1007/978-3-319-46672-9_29

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-46672-9_29

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-46671-2

  • Online ISBN: 978-3-319-46672-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics