Abstract
The training of stacked auto-encoders (SAEs) consists of an unsupervised layer-wise pre-training and a supervised fine-tuning training. The unsupervised pre-training greedily learns internal data representations, initializes network connection weights and brings high generalization. But in the condition that the saliencies of input data are different, the unsupervised pre-training will preserve unimportant contents, lose useful contents and degrade the performance of SAE. In the light of the problem, an iterative stacked weighted auto-encoder (ISWAE) is proposed. To provide robust and discriminative data representations in the unsupervised pre-training, SAE data weighting is embedded in the SAE network by combining the weighting with data reconstruction in an iterative approach. The SAE weights reflect the global saliencies of input data, which are evaluated through transforming implicit weights to explicit weights based on a trained SAE. The weights gained from the current trained SAE are fed into the next SAE through a weighted data reconstruction function. Further, two iterations can yield satisfactory results and an ISWAE is simplified to a two-iteration stacked weighted auto-encoder. Experiments are carried out on MNIST database, CIFAR-10 database and UCI repository. The results show that ISWAE is superior to state-of-art methods: the classification accuracies of ISWAE are in the first place on the chosen data sets; the iterative SAE weighting can be combined with other SAE variant models, producing higher-performance models; it can effectively resolve the problem of data weighting, feasibly to implement without the need of extra coefficients; its computation complexity is controllable, twice that of SAE.





Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Abadi M, Barham P, Chen J, Chen Z, Davis A, Dean J et al. (2016) Tensorflow: a system for large-scale machine learning, In: 12th USENIX symposium on operating systems design and implementation (OSDI ‘16). pp. 265–283.
Battiti R (1994) Using mutual information for selecting features in supervised neural net learning. IEEE Trans Neural Networks 5:537–550
Cament LA, Castillo LE, Perez JP, Galdames FJ, Perez CA (2014) Fusion of local normalization and Gabor entropy weighted features for face identification. Pattern Recogn 47:568–577
Chen M, Zhixiang KQ, Weinberger, Sha F (2013) Marginalized stacked denoising autoencoders. In: Proceeding of the 29th international conference in machine learning, Utah, UT, USA
Chen M, Xu Z, Weinberger K, Fei S (2012) Marginalized denoising autoencoders for domain adaptation. Comput Sci. 767–774
Diakoulaki D, Mavrotas G, Papayannakis L (1995) Determining objective weights in multiple criteria problems: The critic method. Comput Oper Res 22:763–770
Dua D, Karra Taniskidou E (2017) UCI Machine Learning Repository C. U. o. C. Irvine, School of Information and Computer Science, Ed., ed, 2017
Duan B, Pao YH (2006) Iterative feature weighting with neural networks. US Patent US20060224532
Duda RO, Hart P E, Stork DG (2001) Pattern classification.
Gao Z, Shen C, Xie C (2018) Stacked convolutional auto-encoders for single space target image blind deconvolution. Neurocomputing 313:295–305
Geras KJ, Sutton C (2015) Scheduled denoising autoencoders. In: Proceeding of 5th international conference on learning representations. pp. 1–11.
Gilad-Bachrach R, Navot A, Tishby N (2004) Margin based feature selection-theory and algorithms. In: Proceedings of the twenty-first international conference on Machine learning,, p. 43
Hinton GE (2009) Deep belief networks. Scholarpedia 4:5947
Hinton GE, Osindero S, Teh YW (2006) A fast learning algorithm for deep belief nets. Neural Comput 18:1527–1554
Hinton GE, Salakhutdinov RR (2006) Reducing the dimensionality of data with neural networks. Science 313:504–507
Hinton GE, Zemel RS (1993) Autoencoders, minimum description length and Helmholtz free energy. In: International conference on neural information processing systems, Morgan Kaufmann Publishers Inc. pp. 3–10.
Hocke J, Martinetz T (2015) Maximum distance minimization for feature weighting. Pattern Recogn Lett 52:48–52
Kiasari MA, Moirangthem DS, Lee M (2018) Coupled generative adversarial stacked auto-encoder: Cogasa. Neural Netw 100:1–9
Kira K, Rendell LA (1992) A practical approach to feature selection. In: Machine learning proceedings 1992, ed: Elsevier, 1992, pp. 249–256.
Kle´C M (2014) Sparse autoencoders in sentiment analysis. In: Proceeding of 9th international conference on natural language processing, Warsaw
Klys J, Snell J, Zemel R (2018) Learning latent subspaces in variational autoencoders. In: Advances in neural information processing systems. pp. 6444–6454.
Krizhevsky A, Hinton GE (2012) Using very deep autoencoders for content-based image retrieval. In: Proceeding of Esann 2011, European symposium on artificial neural networks, Bruges, Belgium
Krizhevsky A, Hinton G (2009) Learning multiple layers of features from tiny images, Citeseer2009
Le QV (2011) Building high-level features using large scale unsupervised learning. In: IEEE international conference on acoustics, speech and signal processing, Vancouver, BC, Canada. pp. 8595–8598.
LeCun Y, Bottou L, Bengio Y, Haffner P (1998) Gradient-based learning applied to document recognition. Proc IEEE 86:2278–2324
Liu W, Ma T, Tao D, You J (2016) HSAE: a Hessian regularized sparse auto-encoders. Neurocomputing 187:59–65
Masci J, Meier U, Cireşan D, Schmidhuber J (2011) Stacked convolutional auto-encoders for hierarchical feature extraction. In: International conference on artificial neural networks. pp. 52–59
Ortiz A, Ramírez J, Cruz-Arándiga R, García-Tarifa MJ, Martínez-Murcia FJ, Górriz JM (2018) Retinal blood vessel segmentation by multi-channel deep convolutional autoencoder, In: The 13th international conference on soft computing models in industrial and environmental applications. pp. 37–46.
Ruck DW, Rogers SK, Kabrisky M (1990) Feature Selection Using a Multilayer Perceptron. Neural Network Comput 2:40–48
Schölkopf B, Platt J, Hofmann T (2006) Greedy layer-wise training of deep networks. In: International conference on neural information processing systems, Canada. pp. 153–160.
Sun T, Ding S, Li P, Chen W (2019) A comparative study of neural-network feature weighting. Artif Intell Rev 52:469–493
Vincent P, Larochelle H, Lajoie I, Bengio Y, Manzagol PA (2010) Stacked denoising autoencoders: learning useful representations in a deep network with a local denoising criterion. J Mach Learn Research 11:3371–3408
Weston J, Collobert R (2008) Deep learning via semi-supervised embedding. In: International conference on machine learning. pp. 1168–1175.
Xia BY, Bao CC (2013) Speech enhancement with weighted denoising auto-encoder. In: Proc. INTERSPEECH., Lyon, France, pp. 3444–3448
Yingming W (1997) Using the method of maximizing deviation to make decision for multiindices. J Syst Eng Electron 8:21–26
Yu Z-J, Hu X-P, Mao Q (2009) Novel credit rating method under electronic commerce. Control Decis 11:1668–1672
Yu J, Huang D, Wei Z (2017) Unsupervised image segmentation via Stacked denoising auto-encoder and hierarchical patch indexing. Sig Process 143:246–353
Zhai S, Zhang Z.M (2016) Semisupervised autoencoder for sentiment analysis. In: Thirtieth AAAI conference on artificial intelligence. pp. 1394–1400.
Acknowledgements
The work is jointly supported by the National Natural Science Foundation of China (No.61672522 and No.61976216) and the China Postdoctoral Science Foundation (No. 2016M601910).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Additional information
Communicated by V. Loia.
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Sun, T., Ding, S. & Xu, X. An iterative stacked weighted auto-encoder. Soft Comput 25, 4833–4843 (2021). https://doi.org/10.1007/s00500-020-05490-7
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00500-020-05490-7