Skip to main content

Robustness of Raw Images Classifiers Against the Class Imbalance – A Case Study

  • Conference paper
  • First Online:

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 11127))

Abstract

Our aim is to investigate the robustness of classifiers against the class imbalance. From this point of view, we compare several most widely used classifiers as well as the one recently proposed, which is based on the assumption that the probability densities in classes have the matrix normal distribution. As the base for comparison we take a sequence of images from that laser based additive manufacturing process. It is important that the classifiers are fed by raw images. The classifiers are compared according to several criterions and the methodology of all pair-wise comparisons is used to rank them.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. Breiman, L.: Random forests. Mach. Learn. 45, 5–32 (2001)

    Article  Google Scholar 

  2. Devroye, L., Gyorfi, L., Lugosi, G.: A Probabilistic Theory of Pattern Recognition. Springer, Berlin (2013). https://doi.org/10.1007/978-1-4612-0711-5

    Book  MATH  Google Scholar 

  3. Haykin, S.S.: Neural Networks and Learning Machines. Pearson, Upper Saddle River (2009)

    Google Scholar 

  4. Kadziński, M., Słowiński, R.: Parametric evaluation of research units with respect to reference profiles. Decis. Support. Syst. 72, 33–43 (2015)

    Article  Google Scholar 

  5. Krzysko, M., Skorzybut, M., Wolynski, W.: Classifiers for doubly multivariate data. Discuss. Math Probab. Stat., 31 (2011)

    Google Scholar 

  6. Manceur, A.M., Dutilleul, P.: Maximum likelihood estimation for the tensor normal distribution: algorithm, minimum sample size, and empirical bias and dispersion. J. Comput. Appl. Math. 239, 37–49 (2013)

    Article  MathSciNet  Google Scholar 

  7. Ohlson, M., Ahmad, M.R., Von Rosen, D.: The multi-linear normal distribution: introduction and some basic properties. J. Multivar. Anal. 113, 37–47 (2013)

    Article  Google Scholar 

  8. Rafajłowicz, E.: Data structures for pattern and image recognition with application to quality control. Acta Polytechnica Hungarica, Informatics (accepted for publication)

    Google Scholar 

  9. Rafajłowicz, E.: Classifiers for matrix normal images: derivation and testing. In: Rutkowski, L., Scherer, R., Korytkowski, M., Pedrycz, W., Tadeusiewicz, R., Zurada, J.M. (eds.) ICAISC 2018. LNCS (LNAI), vol. 10841, pp. 668–679. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-91253-0_62

    Chapter  Google Scholar 

  10. Rafajłowicz, E., Rafajłowicz, W.: Image-driven decision making with application to control gas burners. In: Saeed, K., Homenda, W., Chaki, R. (eds.) CISIM 2017. LNCS, vol. 10244, pp. 436–446. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-59105-6_37

    Chapter  Google Scholar 

  11. Salvatore, G., Matarazzo, B., Slowinski, R.: Rough approximation of a preference relation by dominance relations. Eur. J. Oper. Res. 117(1), 63–83 (1999)

    Article  Google Scholar 

  12. Schein, A.I., Ungar, L.H.: Active learning for logistic regression: an evaluation. Mach. Learn. 68(3), 235–265 (2007)

    Article  Google Scholar 

  13. Skubalska-Rafajłowicz, E.: Sparse random projections of camera images for monitoring of a combustion process in a gas burner. In: Saeed, K., Homenda, W., Chaki, R. (eds.) CISIM 2017. LNCS, vol. 10244, pp. 447–456. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-59105-6_38

    Chapter  MATH  Google Scholar 

  14. Werner, K., Jansson, M., Stoica, P.: On estimation of covariance matrices with Kronecker product structure. IEEE Trans. Signal Process. 56(2), 478–491 (2008)

    Article  MathSciNet  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ewaryst Rafajłowicz .

Editor information

Editors and Affiliations

A Appendix

A Appendix

The densities of the matrix normal distribution are defined as follows:

$$\begin{aligned} f_j({\mathbf {X}})=\frac{1}{c_j}\,\exp \left[ -\frac{1}{2}\, {\text {tr}}[U_j^{-1}({\mathbf {X}} - {\mathbf {M}}_j)\,V_j^{-1}\, ({\mathbf {X}} - {\mathbf {M}}_j)^T ] \right] , \end{aligned}$$
(18)

where the normalization constants are given by:

$$\begin{aligned} c_j{\mathop {=}\limits ^{def}} (2\, \pi )^{0.5\,{n\, m}}\, {\text {det}}[U_j]^{0.5\,n}\, {\text {det}}[V_j]^{0.5\,m}\, , \end{aligned}$$
(19)

where \(n\times m\) matrices \(M_j\)’s denote the class means matrices. The covariance structure of MND class densities is as follows

  1. 1.

    \(n\times n\) matrix \(U_j\) denotes the covariance matrix between rows of an image from j-th class,

  2. 2.

    \(m\times m\) matrix \(V_j\) stands for the covariance matrix between columns of an image from j-th class.

The above definitions are meaningful only when \({\text {det}}[U_j]>0\), \({\text {det}}[V_j]>0\).

The equivalent description of MND is the following:

$$\begin{aligned} \text {vec}{({\mathbf {X}})} \sim \mathcal {N}_{n\, m}(\text {vec}{({\mathbf {M}}_j}),\, \Sigma _j ), \; {\text {for }} j=1,\, 2,\ldots , J, \end{aligned}$$
(20)

where \(\mathcal {N}_{K}\) stands for the classic (vector valued) normal distribution with K componentnts. In (20), \(\text {vec}{({\mathbf {X}})}\) is the operation of stacking columns of matrix \({\mathbf {X}}\), while \(\Sigma _j\) is a \(n\, m\times n\, m\) covariance matrix of j-th class, which is the Kronecker product (denoted as \(\otimes \)) of \(U_j\) and \(V_j\), i.e.,

$$\begin{aligned} \Sigma _j {\mathop {=}\limits ^{def}} U_j\otimes V_j,\quad j=1,\, 2,\ldots , J. \end{aligned}$$
(21)

Formulas (20) and (21) show clearly that MND’s form a subclass of all normal distributions. Namely, MND’s have the special structure of the covariance matrix given by (21) (see [7]). Thus, in practice, it suffices to estimate two much smaller matrices \(U_j\) and \(V_j\) instead of a general covariance matrix which is \(n\, m\times n\, m\). As the consequence, it suffices to have:

$$\begin{aligned} N_j \ge \max \left\{ \frac{n}{m}, \frac{m}{n} \right\} +1, \quad j=1,\, 2, \ldots ,\, J. \end{aligned}$$
(22)

(see [6] for the proof).

Rights and permissions

Reprints and permissions

Copyright information

© 2018 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Rafajłowicz, E. (2018). Robustness of Raw Images Classifiers Against the Class Imbalance – A Case Study. In: Saeed, K., Homenda, W. (eds) Computer Information Systems and Industrial Management. CISIM 2018. Lecture Notes in Computer Science(), vol 11127. Springer, Cham. https://doi.org/10.1007/978-3-319-99954-8_14

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-99954-8_14

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-99953-1

  • Online ISBN: 978-3-319-99954-8

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics