Abstract
Kernel fisher discriminant analysis (KFDA) has received extensive study in recent years as a dimensionality reduction technique. KFDA always encounters an intrinsic singularity of scatter matrices in the feature space, namely ‘small sample size’ (SSS) problem. Several novel methods have been proposed to cope with this problem. In this paper, kernel uncorrelated discriminant analysis (KUDA) is proposed, which not only can bear on the SSS problem but also extract uncorrelated features, a desirable property for many applications. And then, we have conducted a comparative study on the application of KUDA and other variants of KFDA in radar target recognition problem. The experimental results indicate the effectiveness of KUDA and illustrate the utility of KFDA on the problem.
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Bo, L.F., Wang, L., Jiao, L.C.: Training support vector machines using greedy stagewise algorithm. In: Ho, T.-B., Cheung, D., Liu, H. (eds.) PAKDD 2005. LNCS (LNAI), vol. 3518, pp. 632–638. Springer, Heidelberg (2005)
Bo, L.F., Wang, L., Jiao, L.C.: Feature scaling for kernel fisher discriminant analysis using leave-one-out cross validation. Neural Computation 18(4), 961–978 (2006)
Mika, S., Ratsch, G., Weston, J.: Fisher discriminant analysis with kernels. In: Proceedings of the IEEE Workshop on Neural Networks for signal Processing, pp. 41–48 (1999)
Baudat, G., Anouar, F.: Generalized discriminant analysis using a kernel approach. Nerual Compuation 12(10), 2385–2404 (2000)
Yang, M.H.: Kernel eigenfaces vs. kernel fisherfaces: face recognition using kernel methods. In: Proceedings of Fifth IEEE International Conference Automatic Face and Gesture Recognition, pp. 215–220 (2002)
Park, C.H., Park, H.: Nonlinear discriminant analysis using kernel functions and the generalized singular value decomposition. SIAM Journal on Matrix Analysis and Applications (to appear)
Jin, Z., Yang, J.Y., Tang, Z.M., Hu, Z.S.: A theorem on the uncorrelated optimal discriminant vectors. Pattern Recognition 34, 2041–2047 (2001)
Ye, J.P., Janardan, R., Li, Q., Park, H.: Feature extraction via generalized uncorrelated linear discriminant analysis. In: Proceedings of the 21st International Conference on Machine Learning, Banff, Canada (2004)
Ye, J.P.: Characterization of a family of algorithms for generalized discriminant analysis on undersampled problems. Journal of Machine Learning Research 6(4), 483–502 (2005)
Yang, J., Yang, J.Y.: Why can LDA be performed in PCA transformed space? Pattern Recognition 36, 563–566 (2003)
Howland, P., Park, H.: Generalizing Discriminant Analysis Using the Generalized Singular Value Decomposition. IEEE Transactions on PAMI 26(8), 995–1006 (2004)
Ye, J.P., Janardan, R., Park, C.H., Park, H.: An optimization criterion for generalized discriminant analysis on undersampled problems. IEEE Transactions on PAMI 26(8), 982–994 (2004)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2006 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Wang, L., Bo, L., Jiao, L. (2006). Kernel Uncorrelated Discriminant Analysis for Radar Target Recognition. In: King, I., Wang, J., Chan, LW., Wang, D. (eds) Neural Information Processing. ICONIP 2006. Lecture Notes in Computer Science, vol 4233. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11893257_45
Download citation
DOI: https://doi.org/10.1007/11893257_45
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-46481-5
Online ISBN: 978-3-540-46482-2
eBook Packages: Computer ScienceComputer Science (R0)