Skip to main content

Robust \(L_{2}E\) Parameter Estimation of Gaussian Mixture Models: Comparison with Expectation Maximization

  • Conference paper
  • First Online:
Neural Information Processing (ICONIP 2015)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 9491))

Included in the following conference series:

Abstract

The purpose of this paper is to discuss the use of \(L_{2}E\) estimation that minimizes integrated square distance as a practical robust estimation tool for unsupervised clustering. Comparisons to the expectation maximization (EM) algorithm are made. The \(L_{2}E\) approach for mixture models is particularly useful in the study of big data sets and especially those with a consistent numbers of outliers. The focus is on the comparison of \(L_{2}E\) and EM for parameter estimation of Gaussian Mixture Models. Simulation examples show that the \(L_{2}E\) approach is more robust than EM when there is noise in the data (particularly outliers) and for the case when the underlying probability density function of the data does not match a mixture of Gaussians.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Titterington, D.M., Smith, A.F.M., Markov, U.E.: Statistical Analysis of Finite Mixture Distributions. Wiley, New York (1985)

    MATH  Google Scholar 

  2. McLachlan, G.J., Peel, D.: Finite Mixture Models. Wiley, New York (2000)

    Book  MATH  Google Scholar 

  3. Dempster, A.P., Laird, N.M., Rubin, D.B.: Maximum-likelihood from incomplete data via the EM algorithm. J. Roy. Stat. Soc. Ser. B 39, 1–38 (1977)

    MathSciNet  MATH  Google Scholar 

  4. Aitkin, M., Wilson, G.T.: Mixture models, outliers, and the EM algorithm. Technometrics 22, 325–331 (1980)

    Article  MATH  Google Scholar 

  5. Scott, D.W.: On fitting and adapting of density estimates. Comput. Sci. Stat. 30, 124–133 (1998). (Weisberg, S., ed.)

    Google Scholar 

  6. Scott, D.W.: Remarks on fitting and interpreting mixture models. Comput. Sci. Stat. 31, 104–109 (1999). (Berk, K., Pourahmadi, M., eds.)

    Google Scholar 

  7. Scott, D.W.: Parametric statistical modeling by minimum integrated square error. Technometrics 43, 274–285 (2001)

    Article  MathSciNet  Google Scholar 

  8. Scott, D.W.: Outlier detection and clustering by partial mixture modeling. In: COMPSTAT Symposium. Physica-Verlag/Springer (2004)

    Google Scholar 

  9. Basu, A., Harris, I.R., Hjort, H.L., Jones, M.C.: Robust and efficient estimation by minimizing a density power divergence. Biometrika 85, 549–560 (1998)

    Article  MathSciNet  MATH  Google Scholar 

  10. Markatou, M., Basu, A., Lindsay, B.G.: Weighted likelihood estimating equations with a bootstrap root search. J. Am. Stat. Assoc. 93, 740–750 (1998)

    Article  MathSciNet  MATH  Google Scholar 

  11. Thayasivam, U., Sriram, T.N.: \(L_{2}E\) estimation for mixture complexity for count data. Comput. Stat. Data Anal. 53, 4243–4254 (2009)

    Article  MATH  Google Scholar 

  12. Thayasivam, U., Sriram, T.N., Lee, J.: Simultaneous robust estimation in finite mixtures: the continuous case. J. Indian Stat. Assoc. 50, 277–295 (2012)

    MathSciNet  Google Scholar 

  13. Thayasivam, U., Shetty, S., Kuruwita, C., Ramachandran, R.P.: Detection of anomalies in network traffic using L2E for accurate speaker recognition. In: 55th International Midwest Symposium on Circuits & Systems, Boise, pp. 884–887 (2012)

    Google Scholar 

  14. Kai, Y., Dang, X., Bart, H., Chen, Y.: Robust model-based learning via Spatial-EM algorithm. IEEE Trans. Knowl. Data Eng. 27, 1670–1682 (2015)

    Article  Google Scholar 

  15. Rudemo, M.: Empirical choice of histograms and kernel density estimators. Scand. J. Statist. 9, 65–78 (1982)

    MathSciNet  MATH  Google Scholar 

  16. Bowman, A.W.: An alternative method of cross-validation for the smoothing of density estimates. Biometrika 71, 353–360 (1984)

    Article  MathSciNet  Google Scholar 

  17. R: A Language and Environment for Statistical Computing, R Development Core Team, R Foundation for Statistical Computing, Vienna, Austria, (2011). http://www.R-project.org/

  18. Woodward, W.A., Parr, W.C., Schucany, W.R., Lindsay, H.: A comparison of minimum distance and maximum likelihood estimation of a mixture proportion. J. Am. Stat. Assoc. 79, 590–598 (1984)

    Article  MathSciNet  MATH  Google Scholar 

Download references

Acknowledgment

This work was supported by the National Science Foundation through Grant DUE-1122296.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ravi P. Ramachandran .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2015 Springer International Publishing Switzerland

About this paper

Cite this paper

Thayasivam, U., Kuruwita, C., Ramachandran, R.P. (2015). Robust \(L_{2}E\) Parameter Estimation of Gaussian Mixture Models: Comparison with Expectation Maximization. In: Arik, S., Huang, T., Lai, W., Liu, Q. (eds) Neural Information Processing. ICONIP 2015. Lecture Notes in Computer Science(), vol 9491. Springer, Cham. https://doi.org/10.1007/978-3-319-26555-1_32

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-26555-1_32

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-26554-4

  • Online ISBN: 978-3-319-26555-1

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics