Abstract
Nowadays the use of deep network architectures has become widespread in machine learning. Deep belief networks (DBNs) have deep network architectures to create a powerful generative model using training data. Deep belief networks can be used in classification and feature learning. A DBN can be learned unsupervised, and then the learned features are suitable for a simple classifier (like a linear classifier) with a few labeled data. In addition, according to researches, by using sparsity in DBNs we can learn useful low-level feature representations for unlabeled data. In sparse representation, we have the property that learned features can be interpreted, i.e., correspond to meaningful aspects of the input, and capture factors of variation in the data. Different methods are proposed to build sparse DBNs. In this paper, we proposed a new method that has different behavior according to deviation of the activation of the hidden units from a (low) fixed value. In addition, our proposed regularization term has a variance parameter that can control the force degree of sparseness. According to the results, our new method achieves the best recognition accuracy on the test sets in different datasets with different applications (image, speech and text) and we can achieve incredible results when using a different number of training samples, especially when we have a few samples for training.









Similar content being viewed by others
Notes
Available online at “http://ceit.aut.ac.ir/~keyvanrad/DeeBNetToolbox.html”.
Available online at “http://yann.lecun.com/exdb/mnist/”.
Available online at “https://archive.ics.uci.edu/ml/datasets/ISOLET”.
Available online at “http://qwone.com/~jason/20Newsgroups”.
Available online at “http://qwone.com/~jason/20Newsgroups/20news-bydate-matlab.tgz”.
References
Bai L, Li K, Pei J, Jiang S (2015) Main objects interaction activity recognition in real images. Neural Comput Appl 27:335–348. doi:10.1007/s00521-015-1846-7
Bengio Y (2009) Learning deep architectures for AI. Found Trends Mach Learn 2:1–127
Fanty M, Cole RA (1990) Spoken letter recognition. In: Lippmann RP, Moody JE, Touretzky DS (eds) Proceedings of the 3rd international conference on neural information processing systems (NIPS’90). Morgan Kaufmann Publishers Inc., San Francisco, CA, pp 220–226
Fischer A, Igel C (2014) Training restricted Boltzmann machines: an introduction. Pattern Recognit 47:25–39. doi:10.1016/j.patcog.2013.05.025
Halkias X, Paris S, Glotin H (2013) Sparse penalty in deep belief networks: using the mixed norm constraint. arXiv:1301.3533 [cs, stat]
Hinton G (2010) A practical guide to training restricted Boltzmann machines. Machine Learning Group, University of Toronto
Hinton GE, Osindero S, Teh Y-W (2006) A fast learning algorithm for deep belief nets. Neural Comput 18:1527–1554
Hinton GE, Salakhutdinov RR (2006) Reducing the dimensionality of data with neural networks. Science 313:504–507
Ji N-N, Zhang J-S, Zhang C-X (2014) A sparse-response deep belief network based on rate distortion theory. Pattern Recognit 47:3179–3191
Keyvanrad MA, Homayounpour MM (2015a) Normal sparse deep belief network. In: 2015 international joint conference on neural networks (IJCNN), pp 1–7
Keyvanrad MA, Homayounpour MM (2015b) Deep belief network training improvement using elite samples minimizing free energy. Int J Patt Pattern Artif Intell. doi:10.1142/S0218001415510064
Keyvanrad MA, Homayounpour MM (2014) A brief survey on deep belief networks and introducing a new object oriented toolbox (DeeBNet). arXiv:1408.3264 [cs]
LeCun Y, Cortes C, Burges CJ (1998) The MNIST database of handwritten digits. The MNIST data set is available at http://yann.lecun.com/exdb/mnist/index.html
Lee H (2010) Unsupervised feature learning via sparse hierarchical representations. Ph.D. Thesis, Stanford University
Lee H, Ekanadham C, Ng A (2008) Sparse deep belief net model for visual area V2. Adv Neural Inf Process Syst 20:873–880
Liu Y, Zhou S, Chen Q (2011) Discriminative deep belief networks for visual data classification. Pattern Recognit 44:2287–2296
Marc’Aurelio Ranzato Y, Boureau L, LeCun Y (2007) Sparse feature learning for deep belief networks. Adv Neural Inf Process Syst 20:1185–1192
Mohamed A, Dahl G, Hinton G (2009) Deep belief networks for phone recognition. In: NIPS workshop on deep learning for speech recognition and related applications. Canada, pp 1–9
Murphy KP (2012) Machine learning: a probabilistic perspective. The MIT Press, Cambridge
Nair V (2010) Visual object recognition using generative models of images. Ph.D Thesis, University of Toronto
Nair V, Hinton G (2009) 3D object recognition with deep belief nets. Adv Neural Inf Process Syst 22:1339–1347
Olshausen BA, David JF (1996) Emergence of simple-cell receptive field properties by learning a sparse code for natural images. Nature 381:607–609
Tieleman T (2008) Training restricted Boltzmann machines using approximations to the likelihood gradient. In: Proceedings of the 25th international conference on Machine learning. ACM, New York, NY, USA, pp 1064–1071
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Keyvanrad, M.A., Homayounpour, M.M. Effective sparsity control in deep belief networks using normal regularization term. Knowl Inf Syst 53, 533–550 (2017). https://doi.org/10.1007/s10115-017-1049-x
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10115-017-1049-x