Abstract
There are a lot of noise and redundant information in gene expression data, which will reduce the accuracy of the classification model. Denoising auto encoder can be used to reduce the dimension for high-dimensional gene expression data, and get high-level features with strong classification ability. In order to get a better classification model furtherly, a high-level feature selection method based on information cross-entropy is proposed. Firstly, denoising auto encoder is used to encode high-dimensional original data to get high-level features. Then, the high-level features with low cross entropy are selected to get the low-dimensional mapping of original data, which is used to generate optimized and simplified classification models. The high-level features obtained by the denoising auto encoder can improve the accuracy of the classification model, and the selection of high-level features can improve the generalization ability of the classification model. The classification accuracy of the new method under different Corruption Level values and selection rate are studied experimentally. Experimental results on several gene expression datasets show that the proposed method is effective. Compared with classical and excellent mRMR and SVM-RFE algorithms furtherly, the proposed method shows better accuracy.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Lecun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521(7553), 436 (2015)
Hinton, G.E., Salakhutdinov, R.R.: Reducing the dimensionality of data with neural networks. Science 313(5786), 504–507 (2006)
Bengio, Y., Lamblin, P., Popovici, D., et al.: Greedy layer-wise training of deep networks. In: Advances in Neural Information Processing Systems, pp. 153–160 (2007)
Vincent, P., Larochelle, H., Bengio, Y., et al.: Extracting and composing robust features with denoising autoencoders. In: Proceedings of the 25th International Conference on Machine Learning, pp. 1096–1103. ACM (2008)
Masci, J., Meier, U., Cireşan, D., Schmidhuber, J.: Stacked convolutional auto-encoders for hierarchical feature extraction. In: Honkela, T., Duch, W., Girolami, M., Kaski, S. (eds.) ICANN 2011. LNCS, vol. 6791, pp. 52–59. Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-21735-7_7
Wang, B., Huang, D.S., Jiang, C.J.: A new strategy for protein interface identification using manifold learning method. IEEE Trans. Nanobiosci. 13(2), 118–123 (2014)
Wang, S.L., Li, X.L., Zhang, S.W., et al.: Tumor classification by combining PNN classifier ensemble with neighborhood rough set based gene reduction. Comput. Biol. Med. 40(2), 179–189 (2010)
Gupta, A., Wang, H., Ganapathiraju, M.: Learning structure in gene expression data using deep architectures, with an application to gene clustering. In: IEEE International Conference on Bioinformatics and Biomedicine (BIBM), pp. 1328–1335. IEEE (2015)
Singh, V., Baranwal, N., Sevakula, R.K., et al.: Layerwise feature selection in stacked sparse auto-encoder for tumor type prediction. In: IEEE International Conference on Bioinformatics and Biomedicine (BIBM), pp. 1542–1548. IEEE (2016)
Antoniades, A., Took, C.C.: Speeding up feature selection: a deep-inspired network pruning algorithm. In: International Joint Conference on Neural Networks, pp. 360–366. IEEE (2016)
Srivastava, N., Hinton, G., Krizhevsky, A., et al.: Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15(1), 1929–1958 (2014)
Zhang, Y., Li, S., Wang, T., et al.: Divergence-based feature selection for separate classes. Neurocomputing 101, 32–42 (2013)
Zou, Q., Ni, L., Zhang, T., et al.: Deep learning based feature selection for remote sensing scene classification. IEEE Geosci. Remote Sens. Lett. 12(11), 2321–2325 (2015)
Taherkhani, A., Cosma, G., McGinnity, T.: Deep-FS: a feature selection algorithm for deep boltzmann machines. Neurocomputing 322, 22–37 (2018)
Cai, J., Luo, J., Wang, S., et al.: Feature selection in machine learning: a new perspective. Neurocomputing 300, 70–79 (2018)
Peng, H., Long, F., Ding, C.: Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans. Pattern Anal. Mach. Intell. 27(8), 1226–1238 (2005)
Guyon, I., Weston, J., Barnhill, S., et al.: Gene selection for cancer classification using support vector machines. Mach. Learn. 46(1), 389–422 (2002)
Acknowledgment
The authors would like to acknowledge the assistance provided by National Natural Science Foundation of China (Grant no. 61572180, no. 61472467 and no. 61672011).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Cai, J., Huang, W., Yang, S., Wang, S., Luo, J. (2019). A Selection Method for Denoising Auto Encoder Features Using Cross Entropy. In: Huang, DS., Huang, ZK., Hussain, A. (eds) Intelligent Computing Methodologies. ICIC 2019. Lecture Notes in Computer Science(), vol 11645. Springer, Cham. https://doi.org/10.1007/978-3-030-26766-7_44
Download citation
DOI: https://doi.org/10.1007/978-3-030-26766-7_44
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-26765-0
Online ISBN: 978-3-030-26766-7
eBook Packages: Computer ScienceComputer Science (R0)