Skip to main content
Log in

NBSOM: The naive Bayes self-organizing map

  • Original Article
  • Published:
Neural Computing and Applications Aims and scope Submit manuscript

Abstract

The naive Bayes model has proven to be a simple yet effective model, which is very popular for pattern recognition applications such as data classification and clustering. This paper explores the possibility of using this model for multidimensional data visualization. To achieve this, a new learning algorithm called naive Bayes self-organizing map (NBSOM) is proposed to enable the naive Bayes model to perform topographic mappings. The training is carried out by means of an online expectation maximization algorithm with a self-organizing principle. The proposed method is compared with principal component analysis, self-organizing maps, and generative topographic mapping on two benchmark data sets and a real-world image processing application. Overall, the results show the effectiveness of NBSOM for multidimensional data visualization.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11

Similar content being viewed by others

References

  1. Asuncion A, Newman DJ (2007) UCI machine learning repository. University of California, School of Information and Computer Science, Irvine, CA. http://www.ics.uci.edu/mlearn/MLRepository.html

  2. Bishop CM, Svensén M, Williams CKI (1998) Gtm: The generative topographic mapping. Neural Comput 10:215–234

    Article  Google Scholar 

  3. Celeux G, Govaert G (1995) Gaussian parsimonious clustering models. Pattern Recogn 28:781–793

    Article  Google Scholar 

  4. Chickering DM, Heckerman D (1997) Efficient approximations for the marginal likelihood of bayesian networks with hidden variables. Mach Learn 29:181–212

    Article  MATH  Google Scholar 

  5. Dempster AN, Laird NM, Rubin DB (1977) Maximum likelihood from incomplete data via the em algorithm. J R Statist Soc Ser B 39:1–38

    MathSciNet  MATH  Google Scholar 

  6. Domingos P, Pazzani M (1997) On the optimality of the simple bayesian classifier under zero-one loss. Mach Learn 29:103–130

    Article  MATH  Google Scholar 

  7. Estévez PA, Figueroa CJ (2006) Online data visualization using the neural gas network. Neural Netw 19:923–934

    Article  MATH  Google Scholar 

  8. Friedman J (1997) On bias, variance, 0/1—loss, and the curse-of dimensionality. Data Min Knowl Discov 1:55–77

    Article  Google Scholar 

  9. Van Hulle MM (2002) Joint entropy maximization in kernel-based topographic maps. Neural Comput 14:1887–1906

    Article  MATH  Google Scholar 

  10. Jolliffe IT (1986) Principal component analysis. Springer, New York

    Google Scholar 

  11. Kohonen T (1982) Self-organized formation of topologically correct feature maps. Biol Cybern 43:59–69

    Article  MathSciNet  MATH  Google Scholar 

  12. Kohonen T (1995) Self-organizing maps. Springer, Berlin

    Book  Google Scholar 

  13. Langley P, Sage S (1994) Induction of selective bayesian classifiers. In: Proceedings of the 10th conference uncertainty in artificial intelligence, pp 399–406

  14. Meilă M, Heckerman D (1988) An experimental comparison of several clustering and initialization methods. Microsoft research technical report MSR-TR-98-06

  15. Robbins H, Monro S (1951) A stochastic approximation method. Ann Math Statist 22:400–407

    Article  MathSciNet  MATH  Google Scholar 

  16. Ruz GA, Estévez PA, Perez CA (2005) A neurofuzzy color image segmentation method for wood surface defect detection. For Prod J 55:52–58

    Google Scholar 

  17. Ruz GA, Estévez PA, Ramirez PA (2009) Automated visual inspection system for wood defect classification using computational intelligence techniques. Int J Syst Sci 40:163–172

    Article  MATH  Google Scholar 

  18. Venna J, Kaski S (2001) Neighborhood preservation in nonlinear projection methods: an experimental study. In: Proceedings of ICANN, pp 485–491

  19. Venna J, Kaski S (2005) Local multidimensional scaling with controlled tradeoff between trustworthiness and continuity. In: Proceedings of WSOM’05, 5th workshop on self-organizing maps, pp 695–702

  20. Venna J, Kaski S (2006) Local multidimensional scaling. Neural Netw 19:889–899

    Article  MATH  Google Scholar 

  21. Wu X, Kumar V, Quinlan JR, Ghosh J, Yang Q, Motoda H, McLachlan GJ, Ng A, Liu B, Yu PS, Zhou Z, Steinbach M, Hand DJ, Steinberg D (2007) Top 10 algorithms in data mining. Knowl Inf Syst 14:1–37

    Article  Google Scholar 

  22. Xu L, Xu Y, Chow TWS (2010) Polsom: A new method for multidimensional data visualization. Pattern Recogn 43:1668–1675

    Article  MATH  Google Scholar 

  23. Yin H, Allinson NM (2001) Self-organizing mixture networks for probability density estimation. IEEE Trans Neural Netw 12:405–411

    Article  Google Scholar 

Download references

Acknowledgments

The authors would like to thank the EU-funded I*PROMS Network of Excellence and the ORS Award for financially supporting this research.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Gonzalo A. Ruz.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Ruz, G.A., Pham, D.T. NBSOM: The naive Bayes self-organizing map. Neural Comput & Applic 21, 1319–1330 (2012). https://doi.org/10.1007/s00521-011-0567-9

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00521-011-0567-9

Keywords

Navigation