Skip to main content
Log in

The weights initialization methodology of unsupervised neural networks to improve clustering stability

  • Published:
The Journal of Supercomputing Aims and scope Submit manuscript

Abstract

A study on initialization of connection weights of neural networks is expected to be needed because various deep neural networks based on deep learning have attracted much attention recently. However, studies on the relation between the output value of the active function and the learning performance of the neural network with respect to the connection weight value have been conducted mainly on the supervised learning model. This paper focused on improving the efficiency of autonomous neural network model by studying the connection weight initialization as the neural network model of supervised learning. Adaptive resonance theory (ART) is a major model of autonomous neural network that tries to solve the stability–plasticity dilemma by using bottom-up weights and top-down weights. The conventional weights initialization method of ART was to uniformly set all weights, but the proposed method is to initialize by using pre-trained weights. Experiments show that the ART, which initializes the connectivity weights through the proposed method, performs clustering more reliably.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7

Similar content being viewed by others

References

  1. Abreu T, Amorim AJ, Santos-Junior CR, Lotufo AD, Minussi CR (2018) Multinodal load forecasting for distribution systems using a fuzzy-artmap neural network. Appl Soft Comput 71:307–316

    Article  Google Scholar 

  2. Aggarwal V, Ahlawat AK, Pandey BN (2013) A weight initialization approach for training self organizing maps for clustering applications, In: 3rd IEEE International Conference on Advance Computing Conference (IACC), IEEE, India, pp 1000–1005

  3. Fan W, Watanabe T, Asakura K (2010) Mining underlying correlated-clusters in high-dimensional data streams. Int J Soc Humanist Comput 1(3):282–299

    Article  Google Scholar 

  4. Gavrylenko S, Babenko O, Ignatova E (2018) Development of the disable software reporting system on the basis of the neural network. In: Journal of Physics: Conference Series, vol 998, No. 1. IOP Publishing, p 012009

  5. Glorot X, Bengio Y (2010) Understanding the difficulty of training deep feedforward neural networks. In: Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, Italy, pp 249–256

  6. Grossberg S (2013) Adaptive resonance theory: how a brain learns to consciously attend, learn, and recognize a changing world. Neural Netw 37:1–47

    Article  Google Scholar 

  7. He K, Zhang X, Ren S, Sun J (2015) Delving deep into rectifiers: surpassing human-level performance on imagenet classification. In: Proceedings of the IEEE International Conference on Computer Vision, IEEE, Chile, pp 1026–1034

  8. Hinton GE, Osindero S, Teh YW (2006) A fast learning algorithm for deep belief nets. Neural Comput 18(7):1527–1554

    Article  MathSciNet  Google Scholar 

  9. Kim DH, Cha EY (2003) Hierarchical ann classification model combined with the adaptive searching strategy. J KIISE Softw Appl 30(7):649–658

    Google Scholar 

  10. Kim JS, Kim JH, Kim SJ, Jeon HT (2003) Memory information extension model using adaptive resonance theory, In: International Conference of Korea Institute of Intelligent Systems, pp 652–655

  11. Kim WM, Lee SH (1998) A study of initial weight determination for performance enhancement in backpropagation. In: 1998 Autumn Conference of Korean Information Science Society, Korea, vol 25, No. 2, pp 333–335

  12. Kohonen T (1990) The self-organizing map. Proc IEEE 78(9):1464–1480

    Article  Google Scholar 

  13. Koturwar S, Merchant S (2017) Weight initialization of deep neural networks (DNNs) using data statistics. arXiv preprint arXiv:1710.10570

  14. Mittal M (2018) Energy evaluation of sensor protocols based on artificial neural network approach. In: Editorial Board, p 12

  15. Pacheco AG, Krohling RA, da Silva CA (2018) Restricted Boltzmann machine to determine the input weights for extreme learning machines. Expert Syst Appl 96:77–85

    Article  Google Scholar 

  16. Pan SJ, Qiang Y (2010) A survey on transfer learning. IEEE Trans Knowl Data Eng 22(10):1345–1359

    Article  Google Scholar 

  17. Park GM, Kim JH (2016) Deep adaptive resonance theory for learning biologically inspired episodic memory. In: 2016 International Joint Conference on Neural Networks (IJCNN). IEEE, Canada, pp 5174–5180

  18. Ramos EZ, Nakakuni M, Yfantis E (2017) Quantitative measures to evaluate neural network weight initialization strategies. In: 2017 IEEE 7th Annual Computing and Communication Workshop and Conference (CCWC). IEEE, USA, pp. 1–7

  19. Satpute K, Kumar R (2018) Optimization of adaptive resonance theory neural network using particle swarm optimization technique. In: Advances in Machine Learning and Data Science. Springer, Singapore, pp 1–7

  20. Sutskever I, Hinton GE, Taylor GW (2009) The recurrent temporal restricted boltzmann machine, In: Twenty-Third Annual Conference on Neural Information Processing Systems, Canada, pp 1601–1608

Download references

Acknowledgements

This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (NRF-2018R1D1A1B07049988) and by Next-Generation Information Computing Development Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Science, ICT (NRF-2017M3C4A7083279).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Juntae Kim.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Park, S., Seo, S., Jeong, C. et al. The weights initialization methodology of unsupervised neural networks to improve clustering stability. J Supercomput 76, 6421–6437 (2020). https://doi.org/10.1007/s11227-019-02940-4

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11227-019-02940-4

Keywords

Navigation