Skip to main content
Log in

An iterative stacked weighted auto-encoder

  • Methodologies and Application
  • Published:
Soft Computing Aims and scope Submit manuscript

Abstract

The training of stacked auto-encoders (SAEs) consists of an unsupervised layer-wise pre-training and a supervised fine-tuning training. The unsupervised pre-training greedily learns internal data representations, initializes network connection weights and brings high generalization. But in the condition that the saliencies of input data are different, the unsupervised pre-training will preserve unimportant contents, lose useful contents and degrade the performance of SAE. In the light of the problem, an iterative stacked weighted auto-encoder (ISWAE) is proposed. To provide robust and discriminative data representations in the unsupervised pre-training, SAE data weighting is embedded in the SAE network by combining the weighting with data reconstruction in an iterative approach. The SAE weights reflect the global saliencies of input data, which are evaluated through transforming implicit weights to explicit weights based on a trained SAE. The weights gained from the current trained SAE are fed into the next SAE through a weighted data reconstruction function. Further, two iterations can yield satisfactory results and an ISWAE is simplified to a two-iteration stacked weighted auto-encoder. Experiments are carried out on MNIST database, CIFAR-10 database and UCI repository. The results show that ISWAE is superior to state-of-art methods: the classification accuracies of ISWAE are in the first place on the chosen data sets; the iterative SAE weighting can be combined with other SAE variant models, producing higher-performance models; it can effectively resolve the problem of data weighting, feasibly to implement without the need of extra coefficients; its computation complexity is controllable, twice that of SAE.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5

Similar content being viewed by others

References

  • Abadi M, Barham P, Chen J, Chen Z, Davis A, Dean J et al. (2016) Tensorflow: a system for large-scale machine learning, In: 12th USENIX symposium on operating systems design and implementation (OSDI ‘16). pp. 265–283.

  • Battiti R (1994) Using mutual information for selecting features in supervised neural net learning. IEEE Trans Neural Networks 5:537–550

    Article  Google Scholar 

  • Cament LA, Castillo LE, Perez JP, Galdames FJ, Perez CA (2014) Fusion of local normalization and Gabor entropy weighted features for face identification. Pattern Recogn 47:568–577

    Article  Google Scholar 

  • Chen M, Zhixiang KQ, Weinberger, Sha F (2013) Marginalized stacked denoising autoencoders. In: Proceeding of the 29th international conference in machine learning, Utah, UT, USA

  • Chen M, Xu Z, Weinberger K, Fei S (2012) Marginalized denoising autoencoders for domain adaptation. Comput Sci. 767–774

  • Diakoulaki D, Mavrotas G, Papayannakis L (1995) Determining objective weights in multiple criteria problems: The critic method. Comput Oper Res 22:763–770

    Article  Google Scholar 

  • Dua D, Karra Taniskidou E (2017) UCI Machine Learning Repository C. U. o. C. Irvine, School of Information and Computer Science, Ed., ed, 2017

  • Duan B, Pao YH (2006) Iterative feature weighting with neural networks. US Patent US20060224532

  • Duda RO, Hart P E, Stork DG (2001) Pattern classification.

  • Gao Z, Shen C, Xie C (2018) Stacked convolutional auto-encoders for single space target image blind deconvolution. Neurocomputing 313:295–305

    Article  Google Scholar 

  • Geras KJ, Sutton C (2015) Scheduled denoising autoencoders. In: Proceeding of 5th international conference on learning representations. pp. 1–11.

  • Gilad-Bachrach R, Navot A, Tishby N (2004) Margin based feature selection-theory and algorithms. In: Proceedings of the twenty-first international conference on Machine learning,, p. 43

  • Hinton GE (2009) Deep belief networks. Scholarpedia 4:5947

    Article  Google Scholar 

  • Hinton GE, Osindero S, Teh YW (2006) A fast learning algorithm for deep belief nets. Neural Comput 18:1527–1554

    Article  MathSciNet  Google Scholar 

  • Hinton GE, Salakhutdinov RR (2006) Reducing the dimensionality of data with neural networks. Science 313:504–507

    Article  MathSciNet  Google Scholar 

  • Hinton GE, Zemel RS (1993) Autoencoders, minimum description length and Helmholtz free energy. In: International conference on neural information processing systems, Morgan Kaufmann Publishers Inc. pp. 3–10.

  • Hocke J, Martinetz T (2015) Maximum distance minimization for feature weighting. Pattern Recogn Lett 52:48–52

    Article  Google Scholar 

  • Kiasari MA, Moirangthem DS, Lee M (2018) Coupled generative adversarial stacked auto-encoder: Cogasa. Neural Netw 100:1–9

    Article  Google Scholar 

  • Kira K, Rendell LA (1992) A practical approach to feature selection. In: Machine learning proceedings 1992, ed: Elsevier, 1992, pp. 249–256.

  • Kle´C M (2014) Sparse autoencoders in sentiment analysis. In: Proceeding of 9th international conference on natural language processing, Warsaw

  • Klys J, Snell J, Zemel R (2018) Learning latent subspaces in variational autoencoders. In: Advances in neural information processing systems. pp. 6444–6454.

  • Krizhevsky A, Hinton GE (2012) Using very deep autoencoders for content-based image retrieval. In: Proceeding of Esann 2011, European symposium on artificial neural networks, Bruges, Belgium

  • Krizhevsky A, Hinton G (2009) Learning multiple layers of features from tiny images, Citeseer2009

  • Le QV (2011) Building high-level features using large scale unsupervised learning. In: IEEE international conference on acoustics, speech and signal processing, Vancouver, BC, Canada. pp. 8595–8598.

  • LeCun Y, Bottou L, Bengio Y, Haffner P (1998) Gradient-based learning applied to document recognition. Proc IEEE 86:2278–2324

    Article  Google Scholar 

  • Liu W, Ma T, Tao D, You J (2016) HSAE: a Hessian regularized sparse auto-encoders. Neurocomputing 187:59–65

    Article  Google Scholar 

  • Masci J, Meier U, Cireşan D, Schmidhuber J (2011) Stacked convolutional auto-encoders for hierarchical feature extraction. In: International conference on artificial neural networks. pp. 52–59

  • Ortiz A, Ramírez J, Cruz-Arándiga R, García-Tarifa MJ, Martínez-Murcia FJ, Górriz JM (2018) Retinal blood vessel segmentation by multi-channel deep convolutional autoencoder, In: The 13th international conference on soft computing models in industrial and environmental applications. pp. 37–46.

  • Ruck DW, Rogers SK, Kabrisky M (1990) Feature Selection Using a Multilayer Perceptron. Neural Network Comput 2:40–48

    Google Scholar 

  • Schölkopf B, Platt J, Hofmann T (2006) Greedy layer-wise training of deep networks. In: International conference on neural information processing systems, Canada. pp. 153–160.

  • Sun T, Ding S, Li P, Chen W (2019) A comparative study of neural-network feature weighting. Artif Intell Rev 52:469–493

    Article  Google Scholar 

  • Vincent P, Larochelle H, Lajoie I, Bengio Y, Manzagol PA (2010) Stacked denoising autoencoders: learning useful representations in a deep network with a local denoising criterion. J Mach Learn Research 11:3371–3408

    MathSciNet  MATH  Google Scholar 

  • Weston J, Collobert R (2008) Deep learning via semi-supervised embedding. In: International conference on machine learning. pp. 1168–1175.

  • Xia BY, Bao CC (2013) Speech enhancement with weighted denoising auto-encoder. In: Proc. INTERSPEECH., Lyon, France, pp. 3444–3448

  • Yingming W (1997) Using the method of maximizing deviation to make decision for multiindices. J Syst Eng Electron 8:21–26

    Google Scholar 

  • Yu Z-J, Hu X-P, Mao Q (2009) Novel credit rating method under electronic commerce. Control Decis 11:1668–1672

    Google Scholar 

  • Yu J, Huang D, Wei Z (2017) Unsupervised image segmentation via Stacked denoising auto-encoder and hierarchical patch indexing. Sig Process 143:246–353

    Google Scholar 

  • Zhai S, Zhang Z.M (2016) Semisupervised autoencoder for sentiment analysis. In: Thirtieth AAAI conference on artificial intelligence. pp. 1394–1400.

Download references

Acknowledgements

The work is jointly supported by the National Natural Science Foundation of China (No.61672522 and No.61976216) and the China Postdoctoral Science Foundation (No. 2016M601910).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Shifei Ding.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Additional information

Communicated by V. Loia.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Sun, T., Ding, S. & Xu, X. An iterative stacked weighted auto-encoder. Soft Comput 25, 4833–4843 (2021). https://doi.org/10.1007/s00500-020-05490-7

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00500-020-05490-7

Keywords

Navigation