Abstract
For Multi-label classification, redundant and irrelevant features degrade the performance of classification. To select the best features based on several conflicting objectives, feature selection can be modeled as a large-scale optimization problem. However, most existing multi-objective feature selection methods select the features based on minimizing two well-known objectives, the number of features and classification error, additional objectives can be considered to improve the classification performance. In this study, for the first time, a many-objective optimization method is proposed to select the efficient features for multi-label classification based on not only two mentioned objectives, but also maximizing the correlation between features and labels and minimizing the computational complexity of features. Maximizing the correlation could lead to increasing the accuracy of classification. On the other hand, selecting less complex features decreases the computational complexity of feature extraction phase. The most important aim of this paper is to tackle the multi-label feature selection based on the number of features, classification error, correlation between features and labels, and computational complexity of features, simultaneously. The conducted many-objective feature selection problem is solved using a proposed binary version of NSGA-III algorithm. The binary operator improves the exploration power of optimizer to search the large-scale space. In order to evaluate the proposed algorithm (called binary NSGA-III), a benchmarking experiments is conducted on eight multi-label datasets in terms of several multi-objective assessment metrics, including Hypervolume indicator, Pure Diversity, and Set-coverage. Experimental results show significant improvements for proposed method in comparison with other algorithms.




Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Chen G, Ye D, Xing Z, Chen J, Cambria E (2017) Ensemble application of convolutional and recurrent neural networks for multi-label text categorization. In: International joint conference on neural networks (IJCNN), IEEE, pp 2377–2383
Wang J, Yang Y, Mao J, Huang Z, Huang C, Xu W (2016) CNN-RNN: a unified framework for multi-label image classification. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2285–2294
Li Z, Liao B, Li Y, Liu W, Chen M, Cai L (2018) Gene function prediction based on combining gene ontology hierarchy with multi-instance multi-label learning. RSC Adv 8(50):28503–28509
Ueda N, Saito K (2003) Parametric mixture models for multi-labeled text. In: Advances in neural information processing systems, pp 737–744
Boutell MR, Luo J, Shen X, Brown CM (2004) Learning multi-label scene classification. Pattern Recogn 37(9):1757–1771
Li F, Miao D, Pedrycz W (2017) Granular multi-label feature selection based on mutual information. Pattern Recogn 67:410–423
Lee J, Kim DW (2015) Mutual information-based multi-label feature selection using interaction information. Expert Syst Appl 42(4):2013–2025
Xue B, Zhang M, Browne WN (2013) Particle swarm optimization for feature selection in classification: a multi-objective approach. IEEE Trans Cybern 43(6):1656–1671
Wang CM, Huang YF (2009) Evolutionary-based feature selection approaches with new criteria for data mining: a case study of credit approval data. Expert Syst Appl 36(3):5900–5908
Hagenau M, Liebmann M, Hedwig M, Neumann D (2012) Automated news reading: Stock price prediction based on financial news using context-specific features. In: 45th Hawaii International Conference on System Science (HICSS), IEEE, pp 1040–1049
Yin J, Tao T, Xu J (2015) A multi-label feature selection algorithm based on multi-objective optimization. In: International Joint Conference on Neural Networks (IJCNN), IEEE, pp 1–7
Spolaôr N, Monard MC, Tsoumakas G, Lee HD (2016) A systematic review of multi-label feature selection and a new method based on label construction. Neurocomputing 180:3–15
Xue B, Zhang M, Browne WN, Yao X (2016) A survey on evolutionary computation approaches to feature selection. IEEE Trans Evol Comput 20(4):606–626
Marler RT, Arora JS (2010) The weighted sum method for multi-objective optimization: new insights. Struct Multidiscip Optim 41(6):853–862
Zou D, Li S, Kong X, Ouyang H, Li Z (2019) Solving the combined heat and power economic dispatch problems by an improved genetic algorithm and a new constraint handling strategy. Appl Energy 237:646–670
Zou D, Li S, Kong X, Ouyang H, Li Z (2018) Solving the dynamic economic dispatch by a memory-based global differential evolution and a repair technique of constraint handling. Energy 147:59–80
Shen X, Zou D, Duan N, Zhang Q (2019) An efficient fitness-based differential evolution algorithm and a constraint handling technique for dynamic economic emission dispatch. Energy 186:115801
Taradeh M, Mafarja M, Heidari AA, Faris H, Aljarah I, Mirjalili S, Fujita H (2019) An evolutionary gravitational search-based feature selection. Inf Sci 497:219–239
Hichem H, Elkamel M, Rafik M, Mesaaoud MT, Ouahiba C (2019) A new binary grasshopper optimization algorithm for feature selection problem. J King Saud Univ Comput Inf Sci
Storn R, Price K (1997) Differential evolution-a simple and efficient heuristic for global optimization over continuous spaces. J Glob Optim 11(4):341–359
Rahnamayan S, Wang GG (2008) Solving large scale optimization problems by opposition-based differential evolution (ODE). WSEAS Trans Comput 7(10):1792–1804
Tubishat M, Idris N, Shuib L, Abushariah MA, Mirjalili S (2020) Improved salp swarm algorithm based on opposition based learning and novel local search algorithm for feature selection. Expert Syst Appl 145:113122
Deb K, Jain H (2014) An evolutionary many-objective optimization algorithm using reference-point-based nondominated sorting approach, part i: Solving problems with box constraints. IEEE Trans Evolut Comput 18(4):577–601
Pereira RB, Plastino A, Zadrozny B, Merschmann LH (2016) Categorizing feature selection methods for multi-label classification. In: Artificial intelligence review, pp 1–22
Dendamrongvit S, Vateekul P, Kubat M (2011) Irrelevant attributes and imbalanced classes in multi-label text-categorization domains. Intell Data Anal 15(6):843–859
Wandekokem E, Varejão F, Rauber T (2010) An overproduce-and-choose strategy to create classifier ensembles with tuned SVM parameters applied to real-world fault diagnosis. Progress Pattern Recognit Image Anal Comput Vis Appl, pp 500–508
Spolaor N, Cherman EA, Monard MC, Lee HD (2013) Relieff for multi-label feature selection. In: Brazilian conference on intelligent systems (BRACIS), IEEE, pp 6–11
Reyes O, Morell C, Ventura S (2015) Scalable extensions of the ReliefF algorithm for weighting and selecting features on the multi-label learning context. Neurocomputing 161:168–182
SpolaoR N, Cherman EA, Monard MC, Lee HD (2013) A comparison of multi-label feature selection methods using the problem transformation approach. Electron Notes Theor Comput Sci 292:135–151
Lin Y, Hu Q, Liu J, Chen J, Duan J (2016) Multi-label feature selection based on neighborhood mutual information. Appl Soft Comput 38:244–256
Safavian SR, Landgrebe D (1991) A survey of decision tree classifier methodology. IEEE Trans Syst Man Cybern 21(3):660–674
Zhu P, Xu Q, Hu Q, Zhang C, Zhao H (2018) Multi-label feature selection with missing labels. Pattern Recogn 74:488–502
Jiang L, Yu G, Guo M, Wang J (2020) Feature selection with missing labels based on label compression and local feature correlation. Neurocomputing
Cai Z, Zhu W (2018) Multi-label feature selection via feature manifold learning and sparsity regularization. Int J Mach Learn Cybern 9(8):1321–1334
Sun L, Yin T, Ding W, Xu J (2019) Hybrid multilabel feature selection using bpso and neighborhood rough sets for multilabel neighborhood decision systems. IEEE Access 7:175793–175815
Shi M, Tang Y, Zhu X (2019) Mlne: Multi-label network embedding. IEEE Trans Neural Netw Learn Syst
Zhang Y, Dw Gong, Xy Sun, Yn Guo (2017) A PSO-based multi-objective multi-label feature selection method in classification. Sci Rep 7(1):376
Lee J, Seo W (2018) Kim DW (2018) Effective evolutionary multilabel feature selection under a budget constraint. Complexity
Kashef S, Nezamabadi-pour H (2019) A label-specific multi-label feature selection algorithm based on the pareto dominance concept. Pattern Recogn 88:654–667
Bi J, Periaswamy S, Okada K, Kubota T, Fung G, Salganicoff M, Rao RB (2006) Computer aided detection via asymmetric cascade of sparse hyperplane classifiers. In: Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining, ACM, pp 837–844
Gao D, Vasconcelos N (2005) Integrated learning of saliency, complex features, and object detectors from cluttered scenes. In: IEEE computer society conference on computer vision and pattern recognition, IEEE, vol 2, pp 282–287
Bechikh S, Elarbi M, Said LB (2017) Many-objective optimization using evolutionary algorithms: a survey. In: Recent advances in evolutionary multi-objective optimization, Springer, New York, pp 105–137
Deb K, Pratap A, Agarwal S, Meyarivan T (2002) A fast and elitist multiobjective genetic algorithm: Nsga-ii. IEEE Trans Evol Comput 6(2):182–197
Venkatesh B, Anuradha J (2019) A review of feature selection and its methods. Cybern Inf Technol 19(1):3–26
Ghavamzadeh M, Lazaric A, Maillard O, Munos R (2010) Lstd with random projections. In: Advances in neural information processing systems, pp 721–729
Sorower MS (2010) A literature survey on algorithms for multi-label learning. Oregon State University, Corvallis, p 18
Sun Z, Zhang J, Dai L, Li C, Zhou C, Xin J, Li S (2019) Mutual information based multi-label feature selection via constrained convex optimization. Neurocomputing 329:447–456
Jungjit S, Michaelis M, Freitas AA, Cinatl J (2013) Two extensions to multi-label correlation-based feature selection: a case study in bioinformatics. In: 2013 IEEE international conference on systems, man, and cybernetics, IEEE, pp 1519–1524
Clare A, King RD (2001) Knowledge discovery in multi-label phenotype data. In: European conference on principles of data mining and knowledge discovery, Springer, pp 42–53
Lindeberg T (2012) Scale invariant feature transform
Vidal-Naquet M, Ullman S (2003) Object recognition with informative features and linear classification. ICCV 3:281–288
Mahdavi S, Rahnamayan S, Deb K (2018) Opposition based learning: a literature review. Swarm Evolut Comput 39:1–23
Rahnamayan S, Tizhoosh HR, Salama MM (2008) Opposition-based differential evolution. IEEE Trans Evol Comput 12(1):64–79
Seif Z, Ahmadi MB (2015) Opposition versus randomness in binary spaces. Appl Soft Comput 27:28–37
Mahdavi S, Rahnamayan S, Mahdavi A (2018) Majority voting for discrete population-based optimization algorithms. Soft Comput, pp 1–18
Velasquez M, Hester PT (2013) An analysis of multi-criteria decision making methods. Int J Oper Res 10(2):56–66
Bidgoli AA, Rahnamayan S, Mahdavi S, Deb K (2019) A novel pareto-vikor index for ranking scientists’ publication impacts: a case study on evolutionary computation researchers. In: 2019 IEEE congress on evolutionary computation (CEC), IEEE, pp 2458–2465
Zhang ML, Zhou ZH (2007) ML-KNN: A lazy learning approach to multi-label learning. Pattern Recogn 40(7):2038–2048
Skryjomski P, Krawczyk B, Cano A (2019) Speeding up k-nearest neighbors classifier for large-scale multi-label learning on gpus. Neurocomputing 354:10–19
Tsoumakas G, Katakis I, Vlahavas I (2009) Mining multi-label data. In: Data mining and knowledge discovery handbook, Springer, pp 667–685
Li AD, Xue B, Zhang M (2020) Multi-objective feature selection using hybridization of a genetic algorithm and direct multisearch for key quality characteristic selection. Inf Sci
Bidgoli AA, Ebrahimpour-Komleh H, Rahnamayan S (2020) Reference-point-based multi-objective optimization algorithm with opposition-based voting scheme for multi-label feature selection. Inf Sci 547:1–17
Zitzler E, Thiele L (1998) Multiobjective optimization using evolutionary algorithms—a comparative case study. In: International conference on parallel problem solving from nature, Springer, pp 292–301
While L, Hingston P, Barone L, Huband S (2006) A faster algorithm for calculating hypervolume. IEEE Trans Evol Comput 10(1):29–38
Wang H, Jin Y, Yao X (2017) Diversity assessment in many-objective optimization. IEEE Trans Cybern 47(6):1510–1522
Woolson R (2007) Wilcoxon signed-rank test. Wiley encyclopedia of clinical trials, pp 1–3
Author information
Authors and Affiliations
Corresponding authors
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Asilian Bidgoli, A., Ebrahimpour-komleh, H. & Rahnamayan, S. A novel binary many-objective feature selection algorithm for multi-label data classification. Int. J. Mach. Learn. & Cyber. 12, 2041–2057 (2021). https://doi.org/10.1007/s13042-021-01291-y
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s13042-021-01291-y