Skip to main content

Constraining Type II Error: Building Intentionally Biased Classifiers

  • Conference paper
  • First Online:
Advances in Computational Intelligence (IWANN 2017)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 10306))

Included in the following conference series:

Abstract

In many applications, false positives (type I error) and false negatives (type II) have different impact. In medicine, it is not considered as bad to falsely diagnosticate someone healthy as sick (false positive) as it is to diagnosticate someone sick as healthy (false negative). But we are also willing to accept some rate of false negatives errors in order to make the classification task possible at all. Where the line is drawn is subjective and prone to controversy. Usually, this compromise is given by a cost matrix where an exchange rate between errors is defined. For many reasons, however, it might not be natural to think of this trade-off in terms of relative costs. We explore novel learning paradigms where this trade-off can be given in the form of the amount of false negatives we are willing to tolerate. The classifier then tries to minimize false positives while keeping false negatives within the acceptable bound. Here we consider classifiers based on kernel density estimation, gradient descent modifications and applying a threshold to classifying and ranking scores.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Bessa, S., Domingues, I., Cardoso, J.S., Passarinho, P., Cardoso, P., Rodrigues, V., Lage, F.: Normal breast identification in screening mammography: a study on 18 000 images. In: IEEE International Conference on Bioinformatics and Biomedicine (BIBM), pp. 325–330. IEEE (2014)

    Google Scholar 

  2. Chawla, N.V., Bowyer, K.W., Hall, L.O., Philip Kegelmeyer, W.: SMOTE: synthetic minority over-sampling technique. J. Artif. Intell. Res. 16, 321–357 (2002)

    MATH  Google Scholar 

  3. Domingos, P.: MetaCost: a general method for making classifiers cost-sensitive. In: Proceedings of the Fifth International Conference on Knowledge Discovery, vol. 55, pp. 155–164 (1999)

    Google Scholar 

  4. Gang, W., Chang, E.: Class-boundary alignment for imbalanced dataset learning. Twent. Int. Conf. Mach. Learn. (ICML) 1, 49–56 (2003)

    Google Scholar 

  5. Bach, F.R.: Considering cost asymmetry in learning classifiers. JMLR 7, 1713–1741 (2006)

    MathSciNet  MATH  Google Scholar 

  6. Lichman, M.: UCI Machine Learning Repository (2013)

    Google Scholar 

  7. Mangasarian, O.L., Street, W.N., Wolberg, W.H.: Breast cancer diagnosis and prognosis via linear programming. Oper. Res. 43(4), 570–577 (1995)

    Article  MathSciNet  MATH  Google Scholar 

  8. Cortez, P., Cerdeira, A., Almeida, F., Matos, T., Reis, J.: Modeling wine preferences by data mining from physicochemical properties. Decis. Support Syst. 47(4), 547–553 (2009)

    Article  Google Scholar 

  9. Scott, D.W.: Multivariate Density Estimation. Wiley Series in Probability and Statistics. Wiley, Hoboken (1992)

    Book  MATH  Google Scholar 

  10. Cruz, R., Fernandes, K., Cardoso, J.S., Costa, J.F.P.: Tackling class imbalance with ranking. In: International Joint Conference on Neural Networks (IJCNN). IEEE (2016)

    Google Scholar 

  11. Li, H.: Learning to Rank for Information Retrieval and Natural Language Processing, vol. 4 (2011)

    Google Scholar 

  12. Herbrich, R., Graepel, T., Obermayer, K.: Support vector learning for ordinal regression a risk formulation for ordinal regression. In: Proceedings of the Ninth International Conference on Artificial Neural Networks, pp. 97–102 (1999)

    Google Scholar 

  13. Agarwal, R., Joshi, M.V.: Pnrule: a new framework for learning classifier models in data mining (a case-study in network intrusion detection). In: Proceedings of the SIAM International Conference on Data Mining, pp. 1–17. SIAM (2001)

    Google Scholar 

Download references

Acknowledgment

This work was funded by the Project “NanoSTIMA: Macro-to-Nano Human Sensing: Towards Integrated Multimodal Health Monitoring and Analytics/NORTE-01-0145-FEDER-000016” financed by the North Portugal Regional Operational Programme (NORTE 2020), under the PORTUGAL 2020 Partnership Agreement, and through the European Regional Development Fund (ERDF), and also by Fundação para a Ciência e a Tecnologia (FCT) within PhD grant numbers SFRH/BD/122248/2016 and SFRH/BD/93012/2013.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ricardo Cruz .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2017 Springer International Publishing AG

About this paper

Cite this paper

Cruz, R., Fernandes, K., Pinto Costa, J.F., Cardoso, J.S. (2017). Constraining Type II Error: Building Intentionally Biased Classifiers. In: Rojas, I., Joya, G., Catala, A. (eds) Advances in Computational Intelligence. IWANN 2017. Lecture Notes in Computer Science(), vol 10306. Springer, Cham. https://doi.org/10.1007/978-3-319-59147-6_47

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-59147-6_47

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-59146-9

  • Online ISBN: 978-3-319-59147-6

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics