Abstract
This paper investigates mutual information-based feature selection for high dimensional hyperspectral imagery, which accounts for both the relevance of features on classes and the redundancy among features. A representative method shortly known as min-redundancy and max-relevance (mRMR) was adopted and compared with a baseline method called Max- Relevance (MR) in experiments with AVIRIS hyperspectral data. Supervised classifications were also carried out to identify classification accuracies obtainable with hyperspectral data of reduced dimensionality through five different classifiers. The results confirm that mRMR is more discrimination- informative than MR in feature selection due to the additional redundancy analysis. Different classifiers with different accuracies manifest that a more impact but more informative subset may exist. However, the intrinsic dimensionality which indicates the optimal performance of a classifier remains an issue for further investigation.
Keywords
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Landgrebe, D.: Some Fundamentals and Methods for Hyperspectral Image Data Analysis. In: SPIE International Symposium on Biomedical Optics (Photonics West), San Jose California (1999)
Hsieh, P.F., Landgrebe, D.: Classification of High Dimensional Data. PhD Thesis and School of Electrical & Computer Engineering Technical Report TR-ECE 98-4 (1998)
Luis, J., Landgrebe, D.: Hyperspectral Data Analysis and Feature Reduction Via Projection Pursuit. IEEE Transactions on Geoscience and Remote Sensing 37(6), 2653–2667 (1999)
Jain, A.K., Duin, R.P.W., Mao, J.C.: Statistical Pattern Recognition: a Review. IEEE Transaction on Pattern Analysis and Machine Intelligence 22(1), 4–37 (2000)
Landgrebe, D.: Multispectral Data Analysis: A signal Theory Perspective (2005), http://cobweb.ecn.pudue.edu/~biehl/MultiSpec/Signal_Theory.pdf
Estévez, P.A., Tesmer, M., et al.: Normalized Mutual Information Feature Selection. IEEE Transaction on Neural Networks 20(2), 189–200 (2009)
Ali, E.A., Abdeljalil, E.O., Driss, A.: A Powerful Feature Selection Approach Based on Mutual Information. IJCSNS International Journal of Computer Science and Network Security 8(4), 116–121 (2007)
Battiti, R.: Using Mutual Information for Selecting Features in Supervised Neural Net Learning. IEEE Transactions on Neural Networks 5(4), 537–551 (1994)
Li, Y.F., Xie, M., Goh, T.N.: A Study of Mutual Information Based Feature Selection for Case Based Reasoning in Software Cost Estimation. Expert Systems with Applications (2008)
Oveisi, F., Erfanian, A.: A Minimax Mutual Information Scheme for Supervised Feature Extraction and Its Application to EEG-Based Brain-Computer Interfacing. EURASIP Journal on Advances in Signal Processing 2008, 1–8 (2008)
Kwak, N., Choi, C.H.: Input Feature Selection for Classification Problems. IEEE Transactions on Neural Networks 13(1), 143–160 (2002)
Peng, H.C., Long, F.H., Ding, C.: Feature Selection Based on Mutual Information: Criteria of Max-Dependency, Max-Relevance, and Min-Redundancy. IEEE Transaction on Pattern Analysis and Machine Intelligence 27(8), 1226–1238 (2005)
Colannino, J., Juban, J.: The Best K Measurements Are Not the K Best (2004), http://cgm.cs.mcgill.ca/~athens/cs644/Projects/2004/JustinColannino-JeremieJuban/
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2010 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Yao, N., Lin, Z., Zhang, J. (2010). Feature Selection Based on Mutual Information and Its Application in Hyperspectral Image Classification. In: Bi, Y., Williams, MA. (eds) Knowledge Science, Engineering and Management. KSEM 2010. Lecture Notes in Computer Science(), vol 6291. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-15280-1_52
Download citation
DOI: https://doi.org/10.1007/978-3-642-15280-1_52
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-15279-5
Online ISBN: 978-3-642-15280-1
eBook Packages: Computer ScienceComputer Science (R0)