Abstract
Performance in most pattern classifiers is improved when redundant or irrelevant features are removed, however, this is mainly achieved by high demanding computational methods or successive classifiers construction. This paper shows how Associative Memories can be used to get a mask value which represents a subset of features that clearly identifies irrelevant or redundant information for classification purposes, therefore, classification accuracy is improved while significant computational costs in the learning phase are reduced. An optimal subset of features allows register size optimization, which contributes not only to significant power savings but to a smaller amount of synthesized logic, furthermore, improved hardware architectures are achieved due to functional units size reduction, as a result, it is possible to implement parallel and cascade schemes for pattern classifiers on the same ASIC.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Zavaljevski, N., Stevens, F., Reifman, J.: Support vector machines with selective kernel scaling for protein classification and identification of key amino acid positions. Bioinformatics 18(5), 689–696 (2002)
Blum, A.L., Langley, P.: Selection of relevant features and examples in machine learning. Artificial Intelligence 97(1-2), 245–271 (1997)
Valentini, G., Masulli, F.: Ensembles of Learning Machines. In: Marinaro, M., Tagliaferri, R. (eds.) WIRN 2002. LNCS, vol. 2486, pp. 3–22. Springer, Heidelberg (2002)
Kittler, J., Hatef, M., Duin, R.P.W., Matas, J.: On combining classifiers. IEEE Transactions on Pattern Analysis and Machine Intelligence 20(3), 226–239 (1998)
Lam, L.: Classifier combinations: Implementations and theoretical issues. In: Kittler, J., Roli, F. (eds.) MCS 2000. LNCS, vol. 1857, pp. 77–86. Springer, Heidelberg (2000)
Palm, G., Schwenker, F., Sommer, F.: Neural Associative Memories. In: Palm, G., Schwenker, F., Sommer, F. (eds.) Associative Processing and Processors, pp. 307–326. IEEE Computer Society, Los Alamitos (1997)
Kohonen, T.: Correlation Matrix Memories. IEEE Transactions on Computers 21(4), 353–359 (1972)
Steinbuch, K., Frank, H.: Nichtdigitale Lernmatrizen als Perzeptoren. Kybernetik 1(3), 117–124 (1961)
Anderson, J.A., Rosenfeld, E.: Neurocomputing: Fundations of Research. MIT Press, Cambridge (1990)
Steinbuch, K.: Die Lernmatrix. Kybernetik 1(1), 36–45 (1961)
Aldape-Pérez, M., Yáñez-Márquez, C., López-Leyva, L.O.: Feature Selection Using a Hybrid Associative Classifier with Masking Techniques. In: MICAI 2006, pp. 151–160. IEEE Computer Society, Los Alamitos (2006)
Hassoun, M.H.: Fundamentals of Artificial Neural Networks. MIT Press, Cambridge (1995)
Vasconcelos, N.: Feature selection by maximum marginal diversity: optimality and implications for visual recognition. In: Proceedings IEEE Conference On Computer Vision And Pattern Recognition, vol. 1, pp. 762–769 (2003)
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 2007 Springer Berlin Heidelberg
About this paper
Cite this paper
Aldape-Pérez, M., Yáñez-Márquez, C., Argüelles-Cruz, A.J. (2007). Optimized Associative Memories for Feature Selection. In: Martí, J., Benedí, J.M., Mendonça, A.M., Serrat, J. (eds) Pattern Recognition and Image Analysis. IbPRIA 2007. Lecture Notes in Computer Science, vol 4477. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-72847-4_56
Download citation
DOI: https://doi.org/10.1007/978-3-540-72847-4_56
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-72846-7
Online ISBN: 978-3-540-72847-4
eBook Packages: Computer ScienceComputer Science (R0)