Abstract
Research in AI has proved to be revolutionarily beneficial to humankind from the past few decades. Many supporting techniques have been developed that indirectly evolved AI and directly enhanced various machine learning models; one of them being the feature engineering. It can also be considered as applied-ML. Another is the so-called feature selection which is method in which most contributing feature to final decision making, out of the entire feature space are selected for processing into an ML model. It is not an easy task to precisely calculate the dependency of the output variable onto the candidate features, particularly when the data is high in dimensions. In this regard, this study proposes a novel method named the cisoidal analysis based feature selection (CAFS) which uses both pre-established algorithms as well as a new approach of relating members of feature space to a complex sinusoid (cisoid) mathematically, then using signal processing techniques to eliminate certain elements in the entire feature space for enhanced feature selection and hence to obtain higher classification accuracy. Derived from experiments with five high dimensional datasets, CAFS displays significantly competitive performance than some of the pre-existing algorithms. CAFS is highly advantageous in reducing dimension of feature space in most of the applications.











Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Bach FR (2008) Bolasso: model consistent LASSO estimation through the bootstrap. In: Proceedings of the 25th international conference on machine learning, 5 July 2008, pp 33–40
Battiti R (1994) Using mutual information for selecting features in supervised neural net learning. IEEE Trans Neural Netw 5(4):537–550
Bermingham ML, Pong-Wong R, Spiliopoulou A, Hayward C, Rudan I, Campbell H, Wright AF, Wilson JF, Agakov F, Navarro P, Haley CS (2015) Application of high-dimensional feature selection: evaluation for genomic prediction in man. Sci Rep 5:10312
Brown G, Pocock A, Zhao MJ, Luján M (2012) Conditional likelihood maximisation: a unifying framework for information theoretic feature selection. J Mach Learn Res 13(1):27–66
Forman G (2003) An extensive empirical study of feature selection metrics for text classification. J Mach Learn Res 3:1289–1305
Gao W, Hu L, Zhang P (2018) Class-specific mutual information variation for feature selection. Pattern Recognit 79:328–339
Ghareb AS, Bakar AA, Hamdan AR (2016) Hybrid feature selection based on enhanced genetic algorithm for text categorization. Expert Syst Appl 49:31–47. https://doi.org/10.1016/j.eswa.2015.12.004
Guyon I, Elisseeff A (2003) An introduction to variable and feature selection. J Mach Learn Res 3:1157–1182
Guyon I, Weston J, Barnhill S, Vapnik V (2002) Gene selection for cancer classification using support vector machines. Mach Learn 46(1–3):389–422
Guyon I, Weston J, Barnhill S, Vapnik V (2002) Gene selection for cancer classification using support vector machines. Mach Learn 46(1):389–422. https://doi.org/10.1023/a:1012487302797
Han K, Wang Y, Zhang C, Li C, Xu C (2018) Autoencoder inspired unsupervised feature selection. In: 2018 IEEE international conference on acoustics, speech and signal processing (ICASSP), 15 April 2018. IEEE, pp 2941–2945
Hsu HH, Hsieh CW, Lu MD (2011) Hybrid feature selection by combining filters and wrappers. Expert Syst Appl 38(7):8144–8150. https://doi.org/10.1016/j.eswa.2010.12.156
Jain I, Jain VK, Jain R (2018) Correlation feature selection based improved-binary particle swarm optimization for gene selection and cancer classification. Appl Soft Comput 62:203–215. https://doi.org/10.1016/j.asoc.2017.09.038
Jordan C (1983) Cours d'Analyse de l'École Polytechnique, vol II, Calcul Intégral: Intégrales définies et indéfinies, 2nd edn. Paris
Ke Y, Zhang D (2015) Feature selection and analysis on correlated gas sensor data with recursive feature elimination. Sens Actuators B 212:353–363. https://doi.org/10.1016/j.snb.2015.02.025
Kwak N, Choi CH (2002) Input feature selection by mutual information based on Parzen window. IEEE Trans Pattern Anal Mach Intell 24(12):1667–1671
Lewis DD (1992) Feature selection and feature extract ion for text categorization. In: Speech and natural language: proceedings of a workshop held at Harriman, New York, 23–26 February 1992
Li J, Cheng K, Wang S, Morstatter F, Trevino RP, Tang J, Liu H (2017) Feature selection: a data perspective. ACM Comput Surv 50(6):1–45
Lin D, Tang X (2006) Conditional infomax learning: an integrated framework for feature extraction and fusion. In: European conference on computer vision, 7 May 2006. Springer, Berlin, pp 68–82
Ma J, Teng G (2019) A hybrid multiple feature construction approach for classification using genetic programming. Appl Soft Comput 80:687–699. https://doi.org/10.1016/j.asoc.2019.04.039
Peng H, Long F, Ding C (2005) Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans Pattern Anal Mach Intell 27(8):1226–1238
Saeys Y, Inza I, Larrañaga P (2007) A review of feature selection techniques in bioinformatics. Bioinformatics 23(19):2507–2517
Shannon CE (1948) A mathematical theory of communication. Bell Syst Tech J 27(3):379–423
Soni R, Kumar B, Chand S (2019) Optimal feature and classifier selection for text region classification in natural scene images using Weka tool. Multimed Tools Appl 78:31757–31791. https://doi.org/10.1007/s11042-019-07998-z
Urbanowicz RJ, Meeker M, La Cava W, Olson RS, Moore JH (2018) Relief-based feature selection: introduction and review. J Biomed Inform 85:189–203
Wei G, Zhao J, Feng Y, He A, Yu J (2020) A novel hybrid feature selection method based on dynamic feature importance. Appl Soft Comput 93:106337
Xu J, Tang B, He H, Man H (2016) Semisupervised feature selection based on relevance and redundancy criteria. IEEE Trans Neural Netw Learn Syst 28(9):1974–1984
Yang Y, Pedersen JO (1997) A comparative study on feature selection in text categorization. In: ICML 1997, 8 July 1997, vol 97(412–420), p 35
Zeng Z, Zhang H, Zhang R, Yin C (2015) A novel feature selection method considering feature interaction. Pattern Recognit 48(8):2656–2666
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
Authors declare that they have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Mahajan, S., Pandit, A.K. Hybrid method to supervise feature selection using signal processing and complex algebra techniques. Multimed Tools Appl 82, 8213–8234 (2023). https://doi.org/10.1007/s11042-021-11474-y
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-021-11474-y