skip to main content
10.1145/3205455.3205638acmconferencesArticle/Chapter ViewAbstractPublication PagesgeccoConference Proceedingsconference-collections
research-article

Evolutionary feature subspaces generation for ensemble classification

Authors Info & Claims
Published:02 July 2018Publication History

ABSTRACT

Ensemble learning is a powerful machine learning paradigm which leverages a collection of diverse base learners to achieve better prediction performance than that could be achieved by any individual base learner. This work proposes an evolutionary feature subspaces generation based ensemble learning framework, which formulates the tasks of searching for the most suitable feature subspace for each base learner into a multi-task optimization problem and solve it via an evolutionary multi-task optimizer. Multiple such problems which correspond to different base learners are solved simultaneously via an evolutionary multi-task feature selection algorithm such that solving one problem may help solve some other problems via implicit knowledge transfer. The quality of thus generated feature subspaces is supposed to outperform those obtained by individually seeking the optimal feature subspace for each base learner. We implement the proposed framework by using SVM, KNN, and decision tree as the base learners, proposing a multi-task binary particle swarm optimization algorithm for evolutionary multi-task feature selection, and utilizing the major voting scheme to combine the outputs of the base learners. Experiments on several UCI datasets demonstrate the effectiveness of the proposed method.

References

  1. Hussein Almuallim and Thomas G Dietterich. 1994. Learning boolean concepts in the presence of many irrelevant features. Artificial Intelligence 69, 1-2 (1994), 279--305. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Arthur Asuncion and David Newman. 2007. UCI machine learning repository. (2007).Google ScholarGoogle Scholar
  3. Alberto Bertoni, Raffaella Folgieri, and Giorgio Valentini. 2005. Bio-molecular cancer prediction with random subspace ensembles of support vector machines. Neurocomputing 63 (2005), 535--539. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Tim Blackwell and Jürgen Branke. 2004. Multi-swarm optimization in dynamic environments. In Workshops on Applications of Evolutionary Computation. Springer, 489--500.Google ScholarGoogle ScholarCross RefCross Ref
  5. Robert Bryll, Ricardo Gutierrez-Osuna, and Francis Quek. 2003. Attribute bagging: improving accuracy of classifier ensembles by using random feature subsets. Pattern recognition 36, 6 (2003), 1291--1302.Google ScholarGoogle Scholar
  6. Rohitash Chandra, Abhishek Gupta, Yew-Soon Ong, and Chi-Keong Goh. 2016. Evolutionary multi-task learning for modular training of feedforward neural networks. In International Conference on Neural Information Processing. Springer, 37--46. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Girish Chandrashekar and Ferat Sahin. 2014. A survey on feature selection methods. Computers & Electrical Engineering 40, 1 (2014), 16--28. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Li-Yeh Chuang, Sheng-Wei Tsai, and Cheng-Hong Yang. 2011. Improved binary particle swarm optimization using catfish effect for feature selection. Expert Systems with Applications 38, 10 (2011), 12699--12707. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Rakkrit Duangsoithong and Terry Windeatt. 2010. Bootstrap feature selection for ensemble classifiers. In Industrial Conference on Data Mining. Springer, 28--41. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Richard O Duda, Peter E Hart, David G Stork, et al. 1973. Pattern classification. Vol. 2. Wiley New York.Google ScholarGoogle Scholar
  11. David E Goldberg and John H Holland. 1988. Genetic algorithms and machine learning. Machine learning 3, 2 (1988), 95--99. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. A Gupta, YS Ong, B Da, L Feng, and SD Handoko. 2016. Measuring complementarity between function landscapes in evolutionary multitasking. In 2016 IEEE Congress on Evolutionary Computation, accepted.Google ScholarGoogle Scholar
  13. Abhishek Gupta, Yew-Soon Ong, and Liang Feng. 2016. Multifactorial evolution: toward evolutionary multitasking. IEEE Transactions on Evolutionary Computation 20, 3 (2016), 343--357.Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Isabelle Guyon and André Elisseeff. 2003. An introduction to variable and feature selection. Journal of machine learning research 3, Mar (2003), 1157--1182. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Tin Kam Ho. 1998. The random subspace method for constructing decision forests. IEEE transactions on pattern analysis and machine intelligence 20, 8 (1998), 832--844. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Borhan Kazimipour, Xiaodong Li, and A Kai Qin. 2013. Initialization methods for large scale global optimization. In Evolutionary Computation (CEC), 2013 IEEE Congress on. IEEE, 2750--2757.Google ScholarGoogle ScholarCross RefCross Ref
  17. Borhan Kazimipour, Xiaodong Li, and A Kai Qin. 2014. Effects of population initialization on differential evolution for large scale optimization. In Evolutionary Computation (CEC), 2014 IEEE Congress on. IEEE, 2404--2411.Google ScholarGoogle ScholarCross RefCross Ref
  18. James Kennedy. 2011. Particle swarm optimization. In Encyclopedia of machine learning. Springer, 760--766.Google ScholarGoogle Scholar
  19. Ron Kohavi and George H John. 1997. Wrappers for feature subset selection. Artificial intelligence 97, 1-2 (1997), 273--324. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Igor Kononenko. 1994. Estimating attributes: analysis and extensions of RELIEF. In European conference on machine learning. Springer, 171--182. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. Sotiris Kotsiantis. 2011. Combining bagging, boosting, rotation forest and random subspace methods. Artificial Intelligence Review 35, 3 (2011), 223--240. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Ludmila I Kuncheva and Catrin O Plumpton. 2010. Choosing parameters for random subspace ensembles for fMRI classification. In International Workshop on Multiple Classifier Systems. Springer, 54--63. Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. Ludmila I Kuncheva, Juan J Rodríguez, Catrin O Plumpton, David EJ Linden, and Stephen J Johnston. 2010. Random subspace ensembles for fMRI classification. IEEE transactions on medical imaging 29, 2 (2010), 531--542.Google ScholarGoogle Scholar
  24. Jing J Liang, S Baskar, Ponnuthurai N Suganthan, and A Kai Qin. 2006. Performance evaluation of multiagent genetic algorithm. Natural Computing 5, 1 (2006), 83--96. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Shih-Wei Lin and Shih-Chieh Chen. 2009. PSOLDA: A particle swarm optimization approach for enhancing classification accuracy rate of linear discriminant analysis. Applied Soft Computing 9, 3 (2009), 1008--1015. Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. Thomas Marill and D Green. 1963. On the effectiveness of receptors in recognition systems. IEEE transactions on Information Theory 9, 1 (1963), 11--17. Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. Ahmet Mert, Niyazi Kiliç, and Erdem Bilgili. 2016. Random subspace method with class separability weighting. Expert Systems 33, 3 (2016), 275--285. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Durga Prasad Muni, Nikhil R Pal, and Jyotirmay Das. 2006. Genetic programming for simultaneous feature selection and classifier design. IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics) 36, 1 (2006), 106--117. Google ScholarGoogle ScholarDigital LibraryDigital Library
  29. Luiz S Oliveira, Robert Sabourin, Flávio Bortolozzi, and Ching Y Suen. 2003. Feature selection for ensembles: A hierarchical multi-objective genetic algorithm approach. In Proceedings of the Seventh International Conference on Document Analysis and Recognition-Volume 2. IEEE Computer Society, 676. Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. David W Opitz. 1999. Feature selection for ensembles. AAAI/IAAI 379 (1999), 384. Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. David W Opitz and Richard Maclin. 1999. Popular ensemble methods: An empirical study. J. Artif. Intell. Res.(JAIR) 11 (1999), 169--198. Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. Catrin O Plumpton, Ludmila I Kuncheva, Nikolaas N Oosterhof, and Stephen J Johnston. 2012. Naive random subspace ensemble with linear classifiers for real-time classification of fMRI data. Pattern Recognition 45, 6 (2012), 2101--2108. Google ScholarGoogle ScholarDigital LibraryDigital Library
  33. A Kai Qin and Xiaodong Li. 2013. Differential evolution on the CEC-2013 single-objective continuous optimization testbed. In Evolutionary Computation (CEC), 2013 IEEE Congress on. IEEE, 1099--1106.Google ScholarGoogle Scholar
  34. A Kai Qin, Federico Raimondo, Florence Forbes, and Yew Soon Ong. 2012. An improved CUDA-based implementation of differential evolution on GPU. In Proceedings of the 14th annual conference on Genetic and evolutionary computation. ACM, 991--998. Google ScholarGoogle ScholarDigital LibraryDigital Library
  35. A Kai Qin, Ponnuthurai N Suganthan, and Marco Loog. 2006. Generalized null space uncorrelated Fisher discriminant analysis for linear dimensionality reduction. Pattern Recognition 39, 9 (2006), 1805--1808. Google ScholarGoogle ScholarDigital LibraryDigital Library
  36. Martin Sewell. 2008. Ensemble learning. RN 11, 02 (2008).Google ScholarGoogle Scholar
  37. Yuhui Shi and Russell C Eberhart. 1999. Empirical study of particle swarm optimization. In Evolutionary computation, 1999. CEC 99. Proceedings of the 1999 congress on, Vol. 3. IEEE, 1945--1950.Google ScholarGoogle Scholar
  38. Yisen Wang and Shu-Tao Xia. 2016. A novel feature subspace selection method in random forests for high dimensional data. In Neural Networks (IJCNN), 2016 International Joint Conference on. IEEE, 4383--4389.Google ScholarGoogle ScholarCross RefCross Ref
  39. A Wayne Whitney. 1971. A direct method of nonparametric measurement selection. IEEE Trans. Comput. 100, 9 (1971), 1100--1103. Google ScholarGoogle ScholarDigital LibraryDigital Library
  40. Tsz Ho Wong, A Kai Qin, Shengchun Wang, and Yuhui Shi. 2015. cuSaDE: A CUDA-based parallel self-adaptive differential evolution algorithm. In Proceedings of the 18th Asia Pacific Symposium on Intelligent and Evolutionary Systems-Volume 2. Springer, 375--388.Google ScholarGoogle ScholarCross RefCross Ref
  41. Junshi Xia, Mauro Dalla Mura, Jocelyn Chanussot, Peijun Du, and Xiyan He. 2015. Random subspace ensembles for hyperspectral image classification with extended morphological attribute profiles. IEEE Transactions on Geoscience and Remote Sensing 53, 9 (2015), 4768--4786.Google ScholarGoogle ScholarCross RefCross Ref
  42. Bing Xue, Mengjie Zhang, and Will N Browne. 2014. Particle swarm optimisation for feature selection in classification: Novel initialisation and updating mechanisms. Applied Soft Computing 18 (2014), 261--276. Google ScholarGoogle ScholarDigital LibraryDigital Library
  43. Yunming Ye, Qingyao Wu, Joshua Zhexue Huang, Michael K Ng, and Xutao Li. 2013. Stratified sampling for feature subspace selection in random forests for high dimensional data. Pattern Recognition 46, 3 (2013), 769--787. Google ScholarGoogle ScholarDigital LibraryDigital Library
  44. Shi-Zheng Zhao, Jing J Liang, Ponnuthurai N Suganthan, and Mehmet Fatih Tasgetiren. 2008. Dynamic multi-swarm particle swarm optimizer with local search for large scale global optimization. In Evolutionary Computation, 2008. CEC 2008.(IEEE World Congress on Computational Intelligence). IEEE Congress on. IEEE, 3845--3852.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Evolutionary feature subspaces generation for ensemble classification

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      GECCO '18: Proceedings of the Genetic and Evolutionary Computation Conference
      July 2018
      1578 pages
      ISBN:9781450356183
      DOI:10.1145/3205455

      Copyright © 2018 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 2 July 2018

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article

      Acceptance Rates

      Overall Acceptance Rate1,669of4,410submissions,38%

      Upcoming Conference

      GECCO '24
      Genetic and Evolutionary Computation Conference
      July 14 - 18, 2024
      Melbourne , VIC , Australia

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader