Skip to main content

Unsupervised Parameterisation of Gaussian Mixture Models

  • Conference paper
  • First Online:

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 2504))

Abstract

In this paper we explain a new practical methodology to fully parameterise Gaussian Mixture Models (GMM) to describe data set distributions. Our approach analyses hierarchically a data set distribution to be modeled, determining unsupervisedly an appropriate number of components of the GMM, and their corresponding parameterisation. Results are provided that show the improvement of our method with regard to an implementation of the traditional approach usually applied to solve this problem. The method is also tested in the unsupervised generation of shape models for visual tracking applications.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Cootes, T., Taylor, C.: Statistical models of appearance for computer vision. Technical report, Wolfson Image Analysis Unit. Imaging Science and Biomedical Engineering, University of Manchester (2000)

    Google Scholar 

  2. D. Ponsa, Solé, A., A. López, Cañero, C., Radeva, P., Vitriá, J.: Regularized EM. In Torres, M., Sanfeliu, A., eds.: Pattern Recognition and Applications. Frontiers in Artificial Intelligence and Applications. IOS Press (2000) 69–77

    Google Scholar 

  3. Sekita, I., Kurita, T., Otsu, N., Abdelmalek, N.: A thresholding method using the mixture of normal density functions. In: International Symposium on Speech, Image Processing and Neural Networks, Hong Kong (1994) 304–307

    Google Scholar 

  4. Bishop, C.M.: Neural Networks for Pattern Recognition. Oxford University Press, Oxford (1995)

    Google Scholar 

  5. Dempster, A.P. Laird, N.M., Rubin, D.: Maximum likelihood from incomplete data via the EM algorithm. J. R. Statist. Soc. B 39 (1977) 185–197

    MathSciNet  Google Scholar 

  6. Oliver, J., R.A., B., C.S., W.: Unsupervised Learning using MML. In: Machine Learning: Proceedings of the Thirteenth International Conference (ICML 96), Morgan Kaufmann Publishers, San Francisco, CA (1996) 364–372

    Google Scholar 

  7. Roberts, S., Husmeier, D., Rezek, I., Penny, W.: Bayesian Approaches To Gaussian Mixture Modelling. IEEE Transaction on Pattern Analysis and Machine Intelligence 20 (1998) 1133–1142

    Article  Google Scholar 

  8. Fraley, C., Raftery, A.: How many clusters? which clustering method? answers via model-based cluster analysis. Computer Journal 41 (1998) 578–588

    Article  MATH  Google Scholar 

  9. Bregler, C., Omohundro, S.M.: Surface learning with applications to lipreading. In Cowan, J.D., Tesauro, G., Alspector, J., eds.: Advances in Neural Information Processing Systems. Volume 6., Morgan Kaufmann Publishers, Inc. (1994) 43–50

    Google Scholar 

  10. Richardson, S., Green, P.: On Bayesian analysis of mixtures with an unknown number of components. Journal of the Royal Statistical Society (Series B) 59 (1997) 731–758

    Article  MATH  MathSciNet  Google Scholar 

  11. Roberts, S., Holmes, C., Denison, D.: Minimum entropy data partitioning using Reversible Jump Markov Chain Monte Carlo. IEEE Transactions on Pattern Analysis and Machine Intelligence 23 (2001) 909–915

    Article  Google Scholar 

  12. Friedman, J.H., Rafsky, L.C.: Multivariate generalizations of the Wald-Wolfowitz and Smirnov two-sample tests. Annals of Statistics 7 (1979) 697–717

    Article  MATH  MathSciNet  Google Scholar 

  13. Smith, S.P., Jain, A.K.: A test to determine the multivariate normality of a data set. IEEE Transactions on Pattern Analysis and Machine Intelligence 10 (1988) 757–761

    Article  Google Scholar 

  14. Schwarz, G.: Estimating the dimension of a model. Annals of Statistics 6 (1978) 461–464

    Article  MATH  MathSciNet  Google Scholar 

  15. Heap, T., Hogg, D.: Wormholes in shape space: Tracking through discontinuous changes in shape. In: Sixth International Conference on Computer Vision. (1998) 344–349

    Google Scholar 

  16. Cootes, T., Taylor, C.: A mixture model for representing shape variation. Image and Vision Computing 17 (1999) 567–574

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2002 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Ponsa, D., Roca, X. (2002). Unsupervised Parameterisation of Gaussian Mixture Models. In: Escrig, M.T., Toledo, F., Golobardes, E. (eds) Topics in Artificial Intelligence. CCIA 2002. Lecture Notes in Computer Science(), vol 2504. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-36079-4_34

Download citation

  • DOI: https://doi.org/10.1007/3-540-36079-4_34

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-00011-2

  • Online ISBN: 978-3-540-36079-7

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics