Abstract
The k nearest neighbor is a lazy learning algorithm that is inefficient in the classification phase because it needs to compare the query sample with all training samples. A template reduction method is recently proposed that uses only samples near the decision boundary for classification and removes those far from the decision boundary. However, when class distributions overlap, more border samples are retrained and it leads to inefficient performance in the classification phase. Because the number of reduced samples are limited, using an appropriate feature reduction method seems a logical choice to improve classification time. This paper proposes a new prototype reduction method for the k nearest neighbor algorithm, and it is based on template reduction and ViSOM. The potential property of ViSOM is displaying the topology of data on a two-dimensional feature map, it provides an intuitive way for users to observe and analyze data. An efficient classification framework is then presented, which combines the feature reduction method and the prototype selection algorithm. It needs a very small data size for classification while keeping recognition rate. In the experiments, both of synthetic and real datasets are used to evaluate the performance. Experimental results demonstrate that the proposed method obtains above 70 % speedup ratio and 90 % compression ratio while maintaining similar performance to kNN.
Similar content being viewed by others
References
Hart PE, Stock DG, Duda RO (2001) Pattern classification, 2nd edn. Wiley, Hoboken
Mahmoud SA, Al-Khatib WG (2011) Recognition of Arabic (Indian) bank check digits using log-Gabor filters. Appl Intell 35(3):445–456
Zhao L, Wang L, Xu Q (2012) Data stream classification with artificial endocrine system. Appl Intell 37(3):390–404
Malek H, Ebadzadeh MM, Rahmati M (2012) Three new fuzzy neural networks learning algorithms based on clustering, training error and genetic algorithm. Appl Intell 37(3):280–289
Chen Y, Garcia E, Gupta M, Rahimi A, Cazzanti L (2009) Similarity based classification: concepts and algorithms. J Mach Learn Res 10:747–776
Domeniconi C, Jing P, Gunopulos D (2002) Locally adaptive metric nearest neighbor classification. IEEE Trans Pattern Anal Mach Intell 24(9):1281–1285
Paredes R, Vidal E (2006) Leaning weighted metric to minimize nearest-neighbor classification error. IEEE Trans Pattern Anal Mach Intell 28(7):1100–1110
Wang J, Neskovic P, Cooper L (2007) Improving nearest neighbor rule with a simple adaptive distance measure. Pattern Recognit Lett 28(2):207–213
Li BY, Chen YW, Chen YQ (2008) The nearest neighbor algorithm of local probability centers. IEEE Trans Syst Man Cybern, Part B, Cybern 38(1):141–154
Hart PE (1968) The condensed nearest neighbor rule. IEEE Trans Inf Theory 14(3):515–516
Wilson DR, Martinez TR (2000) Reduction techniques for instance-based learning algorithms. Mach Learn 38(3):257–286
Angiulli F (2007) Fast nearest neighbor condensation for large data sets classification. IEEE Trans Knowl Data Eng 19(11):1450–1464
Pekalska E, Duin RPW, Paclık P (2006) Prototype selection for dissimilarity based classifiers. Pattern Recognit 39(2):189–208
Wu Y, Ianakiev K, Govindaraju V (2002) Improved k-nearest neighbor classification. Pattern Recognit 35:2311–2318
Chang CL (1974) Finding prototypes for nearest neighbor classifiers. IEEE Trans Comput C-23(11):1179–1184
Veenman CJ, Reinders MJT (2005) The nearest subclass classifier: a compromise between the nearest mean and nearest neighbor classifier. IEEE Trans Pattern Anal Mach Intell 27(9):1417–1429
Kohonen T (1990) The self-organizing map. Proc IEEE 78(9):1464–1480
Lam W, Keung CK, Liu D (2002) Discovering useful concept prototypes for classification based on filtering and abstraction. IEEE Trans Pattern Anal Mach Intell 14(8):1075–1090
Nanni L, Lumini A (2008) Particle swarm optimization for prototype reduction. Neurocomputing 72(4–6):1092–1097
Garcia S, Derrac J, Cano JR, Herrera F (2012) Prototype selection for nearest neighbor classification: taxonomy and empirical study. IEEE Trans Pattern Anal Mach Intell 34(3):417–435
Triguero I, Derrac J, Garcıa S, Herrera F (2012) A taxonomy and experimental study on prototype generation for nearest neighbor classification. IEEE Trans Syst Man Cybern, Part C, Appl Rev 42(1):86–100
Fayed HA, Atiya AF (2009) A novel template reduction approach for the k-nearest neighbor method. IEEE Trans Neural Netw 20(5):890–896
Bohlooli A, Jamshidi K (2012) A GPS-free method for vehicle future movement directions prediction using SOM for VANET. Appl Intell 36(3):685–697
Wu J, Li IJ (2010) A SOM-based dimensionality reduction method for KNN classifiers. In: International Conference on System Science and Engineering, pp 173–178
Yin H (2002) ViSOM-a novel method for multivariate data projection and structure visualization. IEEE Trans Neural Netw 13(1)
Wilson DL (1972) Asymptotic properties of nearest neighbor rules using edited data. IEEE Trans Syst Man Cybern 2(3):408–421
Tomek I (1976) An experiment with the edited nearest-neighbor rule. IEEE Trans Syst Man Cybern 6(6):448–452
Sanchez JS, Barandela R, Marques AI, Alejo R, Badenas J (2003) Analysis of new techniques to obtain quality training sets. Pattern Recognit Lett 24(7):1015–1022
Aha DW, Kibler D, Albert MK (1991) Instance-based learning algorithms. Mach Learn 6(1):37–66
Ho SY, Liu CC, Liu S (2002) Design of an optimal nearest neighbor classifier using an intelligent genetic algorithm. Pattern Recognit Lett 23(13):1495–1503
Garcia S, Cano JR, Herrera F (2008) A memetic algorithm for evolutionary prototype selection: a scaling up approach. Pattern Recognit 41(8):2693–2709
Gates W (1972) The reduced nearest neighbor rule. IEEE Trans Inf Theory 18(3):431–433
Devi FS, Murty MN (2002) An incremental prototype set building technique. Pattern Recognit 35(2):505–513
Berglund E (2010) Improved PLSOM algorithm. Appl Intell 32(1):122–130
Kamimura R (2011) Structural enhanced information and its application to improved visualization of self-organizing maps. Appl Intell 34(1):102–115
Theodoridis S, Koutroumbas K (2006) Pattern recognition, 3rd edn. Academic Press, San Diego
Blake C, Keogh E, Merz CJ (2009) UCI repository of machine learning databases. Department of Information and Computer Science, University of California. http://www.ics.uci.edu/~mlearn
Kohonen T (1993) Things you haven’t heard about the self-organizing map. In: IEEE International Conference on Neural Networks, vol 3, pp 1147–pages 1156
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Li, IJ., Chen, JC. & Wu, JL. A fast prototype reduction method based on template reduction and visualization-induced self-organizing map for nearest neighbor algorithm. Appl Intell 39, 564–582 (2013). https://doi.org/10.1007/s10489-013-0433-9
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-013-0433-9