Abstract
In this paper, we propose a novel Deep Auto-Encoders Ensemble model (DAEE) through assembling multiple deep network models with different activation functions. The hidden features obtained by our proposed model have better robustness in representation than traditional variants of auto-encoders because it aggregates the diversified feature representations from multiple activation sub-networks into a more robust uniform feature representations. In order to obtain such uniform feature representations, we set the weight of each individual auto-encoder sub-network by optimizing the cost function over multiple auto-encoder sub-networks in the proposed model. Therefore, decreases the influence of individual sub-networks with improper activations and increase those with appropriate activations, to ensure the final feature representation to keep more predominant and comprehensive feature information. Extensive experiments on benchmark computer vision datasets, including MNIST, COIL-20, CIFAR-10 and SVHN, demonstrate the superiority of our proposed method among state-of-the-art auto-encoder methods, such as sparse auto-encoders (SAE), denoising auto-encoders (DAE), stacked denoising auto-encoders (SDAE) and graph regularized auto-encoders (GAE).
Similar content being viewed by others
References
Mm A, Mas A, Aml A, Rs B, Yu D, A noise robust convolutional neural network for image classification, Results in Engineering https://doi.org/10.1016/j.rineng.2021.100225.
Weninger F, Watanabe S, Tachioka Y, Schuller B (2014) Deep recurrent de-noising auto-encoder and blind de-reverberation for reverberated speech recognition, In: IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), IEEE, 2014, pp 4623–4627
Shahana P, Omman B (2015) Evaluation of features on sentimental analysis. Procedia Computer Science 46:1585–1592
Segal-Rozenhaimer M, Li A, Das K, Chirayath V (2020) Cloud detection algorithm for multi-modal satellite imagery using convolutional neural-networks (cnn). Remote Sensing of Environment 237:111446
Chen C, Zhuo R, Ren J (2019) Gated recurrent neural network with sentimental relations for sentiment classification. Information Sciences 502:268–278
Lin D, Xu G, Xu W, Wang Y, Sun X, Fu K (2020) Scrsr: An efficient recursive convolutional neural network for fast and accurate image super-resolution. Neurocomputing 398:399–407
Garcia KD, de Sá CR, Poel M, Carvalho T, Mendes-Moreira J, Cardoso JM, de Carvalho AC, Kok JN (2021) An ensemble of autonomous auto-encoders for human activity recognition. Neurocomputing 439:271–280
Shen X-J, Ni C, Wang L, Zha Z-J (2021) Sliker: Sparse loss induced kernel ensemble regression. Pattern Recognition 109:107587
Shen X-J, Liu S-X, Bao B-K, Pan C-H, Zha Z-J, Fan J (2020) A generalized least-squares approach regularized with graph embedding for dimensionality reduction. Pattern Recognition 98:107023
Zhang Y, Lu Z, Wang S (2021) Unsupervised feature selection via transformed auto-encoder. Knowledge-Based Systems 215:106748
Yu J, Liu G (2021) Extracting and inserting knowledge into stacked denoising auto-encoders. Neural Networks 137:31–42
Sun W, Shao S, Zhao R, Yan R, Zhang X, Chen X (2016) A sparse auto-encoder-based deep neural network approach for induction motor faults classification. Measurement 89:171–178
Bilgili E, Göknar İC, Ucan ON (2005) Cellular neural network with trapezoidal activation function. International journal of circuit theory and applications 33(5):393–417
Zhang C, Woodland PC (2016) Dnn speaker adaptation using parameterised sigmoid and relu hidden activation functions, In: IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), IEEE, 2016, pp 5300–5304
Zhang C, Woodland PC (2016) Dnn speaker adaptation using parameterised sigmoid and relu hidden activation functions, In: IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), IEEE, 2016, pp 5300–5304
Ward JH Jr (1963) Hierarchical grouping to optimize an objective function. Journal of the American statistical association 58(301):236–244
Han J, Moraga C (1995) The influence of the sigmoid function parameters on the speed of backpropagation learning, In: International workshop on artificial neural networks, Springer, 1995, pp 195–201
Bergstra J, Desjardins G, Lamblin P, Bengio Y, Quadratic polynomials learn better image features, Technical report, 1337
Clevert D-A, Unterthiner T, Hochreiter S, Fast and accurate deep network learning by exponential linear units (elus), arXiv:arXiv:1511.07289
Scardapane S, Scarpiniti M, Comminiello D, Uncini A (2017) Learning activation functions from data using cubic spline interpolation, In: Italian Workshop on Neural Nets, Springer, 2017, pp 73–83
Goodfellow I, Warde-Farley D, Mirza M, Courville A, Bengio Y (2013) Maxout networks, In: International conference on machine learning, PMLR, 2013, pp 1319–1327
Chen J, Combinatorially generated piecewise activation functions, arXiv:arXiv:1605.05216
Harmon M, Klabjan D, Activation ensembles for deep neural networks, arXiv:arXiv:1702.07790
Wen J, Han N, Fang X, Fei L, Yan K, Zhan S (2018) Low-rank preserving projection via graph regularized reconstruction. IEEE Transactions on Cybernetics 49(4):1279–1291
LeCun Y, Bottou L, Bengio Y, Haffner P (1998) Gradient-based learning applied to document recognition. Proceedings of the IEEE 86(11):2278–2324
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition, In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016
Huang G, Liu Z, van der Maaten L, Weinberger KQ (2017) Densely connected convolutional networks, In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017
Gubbi J, Buyya R, Marusic S, Palaniswami M (2013) Internet of things (iot): A vision, architectural elements, and future directions. Future generation computer systems 29(7):1645–1660
Zhang Y, Zhang E, Chen W (2016) Deep neural network for halftone image classification based on sparse auto-encoder. Engineering Applications of Artificial Intelligence 50:245–255
Vincent P, Larochelle H, Bengio Y, Manzagol P-A (2008) Extracting and composing robust features with denoising autoencoders, in: Proceedings of the 25th international conference on Machine learning, 2008, pp 1096–1103
Majumdar A (2018) Blind denoising autoencoder. IEEE transactions on neural networks and learning systems 30(1):312–317
Witten IH, Frank E (2002) Data mining: practical machine learning tools and techniques with java implementations. Acm Sigmod Record 31(1):76–77
Badem H, Caliskan A, Basturk A, Yuksel ME (2016) Classification of human activity by using a stacked autoencoder in medical technologies national congress (TIPTEKNO). IEEE 2016:1–4
Zhang, Y Liu R, Zhang S, Zhu M (2013) Occlusion-robust face recognition using iterative stacked denoising autoencoder, in: International Conference on Neural Information Processing, Springer, 2013, pp 352–359
Budiman, A Fanany MI, Basaruddin C (2014) Stacked denoising autoencoder for feature representation learning in pose-based action recognition, in: 2014 IEEE 3rd Global Conference on Consumer Electronics (GCCE), IEEE, 2014, pp 684–688
Liao Y, Wang Y, Liu Y (2016) Graph regularized auto-encoders for image representation. IEEE Transactions on Image Processing 26(6):2839–2852
Roweis ST (2000) Saul LK. Nonlinear dimensionality reduction by locally linear embedding, science 290(5500):2323–2326
Lu J, Tan Y-P (2009) Regularized locality preserving projections and its extensions for face recognition, IEEE Transactions on Systems, Man, and Cybernetics. Part B (Cybernetics) 40(3):958–963
Zhang X, Zhu Q, Jiang Z-Y, He Y, Xu Y (2018) A novel ensemble model using plsr integrated with multiple activation functions based elm: Applications to soft sensor development. Chemometrics and Intelligent Laboratory Systems 183:147–157
Zhang X-H, Zhu Q-X, He Y-L, Xu Y (2018) A novel robust ensemble model integrated extreme learning machine with multi-activation functions for energy modeling and analysis: Application to petrochemical industry. Energy 162:593–602
Lauly S, Larochelle H, Khapra MM, Ravindran B, Raykar V, Saha A, et al, An autoencoder approach to learning bilingual word representations, arXiv:arXiv:1402.1454
Zhang L, Zhang Q, Zhang L, Tao D, Huang X, Du B (2015) Ensemble manifold regularized sparse low-rank approximation for multiview feature embedding. Pattern Recognition 48(10):3102–3112
Wang M, Hua X-S, Hong R, Tang J, Qi G-J, Song Y (2009) Unified video annotation via multigraph learning. IEEE Transactions on Circuits and Systems for Video Technology 19(5):733–746
LeCun Y, Bottou L, Orr GB, Müller K-R et al (1998) Neural networks: Tricks of the trade. Springer Lecture Notes in Computer Sciences 1524(5–50):6
Jia K, Sun L, Gao S, Song Z, Shi BE (2015) Laplacian auto-encoders: An explicit learning of nonlinear data manifold. Neurocomputing 160:250–260
Acknowledgements
This work was funded in part by the National Natural Science Foundation of China (No.61572240, 61772244).
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
This work was funded in part by the National Natural Science Foundation of China (No. 61572240, 61806086)
Rights and permissions
About this article
Cite this article
Qiang, N., Shen, XJ., Huang, CB. et al. Diversified feature representation via deep auto-encoder ensemble through multiple activation functions. Appl Intell 52, 10591–10603 (2022). https://doi.org/10.1007/s10489-021-03054-2
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-021-03054-2