Abstract
Feature selection is an effective technique in dealing with dimensionality reduction for classification task, a main component of data mining. It searches for an “optimal” subset of features. The search strategies under consideration are one of the three: complete, heuristic, and probabilistic. Existing algorithms adopt various measures to evaluate the goodness of feature subsets. This work focuses on one measure called consistency. We study its properties in comparison with other major measures and different ways of using this measure in search of feature subsets. We conduct an empirical study to examine the pros and cons of these different search methods using consistency. Through this extensive exercise, we aim to provide a comprehensive view of this measure and its relations with other measures and a guideline of the use of this measure with different search strategies facing a new application.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
H. Almuallim and T. G. Dietterich. Learning boolean concepts in the presence of many irrelevant features. Artificial Intelligence, 69(1–2):279–305, November 1994.
M. Ben-Bassat. Pattern recognition and reduction of dimensionality. In P. R. Krishnaiah and L. N. Kanal, editors, Handbook of Statistics, pages 773–791. North Holland, 1982.
A. L. Blum and P. Langley. Selection of relevant features and examples in machine learning. Artificial Intelligence, 97:245–271, 1997.
A Blumer, A. Ehrenfeucht, D. Haussler, and M. K. Warmuth. Occam’s razor. Readings in Machine Learning, pages 201–204, 1990.
G. Brassard and P. Bratley. Fundamentals of Algorithms. Prentice Hall, New Jersy, 1996.
M. Dash. Feature selection via set cover. In Proceedings of IEEE Knowledge and Data Engineering Exchange Eorkshop, pages 165–171, Newport, California, November 1997. IEEE Computer Society.
M. Dash and H. Liu. Feature selection methods for classification. Intelligent Data Analysis: An Interbational Journal, 1(3), 1997.
P. A. Devijver and J. Kittler. Pattern Recognition: A Statistical Approach. Prentice Hall, 1982.
G. H. John, R. Kohavi, and K. Pfleger. Irrelevant features and the subset selection problem. In Proceedings of the Eleventh International Conference on Machine Learning, pages 121–129, 1994.
D. S. Johnson. Approximation algorithms for combinatorial problems. Journal of Computer and System Sciences, 9:256–278, 1974.
K. Kira and L. A. Rendell. The feature selection problem: Traditional methods and a new algorithm. In Proceedings of Ninth National Conference on AI, pages 129–134, 1992.
R. Kohavi. Wrappers for performance enhancement and oblivious decision graphs. PhD thesis, Department of Computer Science, Stanford University, CA, 1995.
H. Liu, H. Motoda, and M. Dash. A monotonic measure for optimal feature selection. In Proceedings of European Conference on Machine Learning, pages 101–106, 1998.
H. Liu and R. Setiono. Feature selection and classification-a probabilistic wrapper approach. In Proceedings of Ninth International Conference on Industrial and Engineering Applications of AI and ES, 1996.
C. J. Merz and P. M. Murphy. UCI repository of machine learning databases, 1996. FTP from ics.uci.edu in the directory pub/machine-learning-databases.
P. M. Narendra and K. Fukunaga. A branch and bound algorithm for feature selection. IEEE Transactions on Computers, C-26(9):917–922, September 1977.
J. R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kaufmann, San Mateo, California, 1993.
T. W. Rauber. Inductive Pattern Classification Methods-Features-Sensors. PhD thesis, Department of Electrical Engineering, Universidale Nova de Lisboa, 1994.
W. Siedlecki and J Sklansky. On automatic feature selection. International Journal of Pattern Recognition and Artificial Intelligence, 2:197–220, 1988.
S. Watanabe. Pattern Recognition: Human and Mechanical. Wiley Intersceince, 1985.
A. Zell and et al. Stuttgart Neural Network Simulator (SNNS), user manual, version 4.1. Technical report, 1995.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2000 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Dash, M., Liu, H., Motoda, H. (2000). Consistency Based Feature Selection. In: Terano, T., Liu, H., Chen, A.L.P. (eds) Knowledge Discovery and Data Mining. Current Issues and New Applications. PAKDD 2000. Lecture Notes in Computer Science(), vol 1805. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-45571-X_12
Download citation
DOI: https://doi.org/10.1007/3-540-45571-X_12
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-67382-8
Online ISBN: 978-3-540-45571-4
eBook Packages: Springer Book Archive