Abstract
The nearest neighbor classification is a simple and effective technique for pattern recognition. The performance of this technique is known to be sensitive to the distance function used in classifying a test instance. In this paper, we propose a technique to learn sample weights via maximizing classification consistency. Experimental analysis shows that the distance trained in this way enlarges the classification consistency on several datasets and has a strong ability to tolerate noise. Moreover, the proposed approach has better performance than nearest neighbor classification and several state-of-the-art methods.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Yao, Y., Zhao, Y.: Attribute reduction in decision-theoretic rough set models. Information Sciences 78(17), 3356–3373 (2008)
Hart, P., Cover, T.: Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 13, 21–27 (1967)
Gilad-Bachrach, R., Navot, A., Tishby, N.: Margin based feature selection - theory and algorithms. In: ICML 2004 (2004)
Wang, J., Neskovic, P., Cooper, L.N.: Improving nearest neighbor rule with a simple adaptive distance measure. Pattern Recognition Letters 28, 207–213 (2007)
Weinberger, K., Blitzer, J., Saul, L.: Distance metric learning for large margin nearest neighbor classification. In: Advances in Neural Information Processing Systems (NIPS), vol. 18
Paredes, R., Vidal, E.: Learning weighted metrics to minimize nearest-neighbor classification error. IEEE Trans. Pattern Anal. Mach. Intell. 28, 1100–1114 (2006)
Hastie, T., Tibshirani, R.: Discriminant Adaptive Nearest Neighbor Classification and Regression. In: Advances in Neural Information Processing Systems, vol. 8, pp. 409–415 (1996)
Howe, N., Cardie, C.: Examining Locally Varying Weights for Nearest Neighbor Algorithms. In: Leake, D.B., Plaza, E. (eds.) ICCBR 1997. LNCS, vol. 1266, pp. 455–466. Springer, Heidelberg (1997)
Kohavi, R., Langley, P., Yung, Y.: The Utility of Feature Weighting in Nearest-Neighbor Algorithms. In: van Someren, M., Widmer, G. (eds.) ECML 1997. LNCS, vol. 1224, pp. 455–466. Springer, Heidelberg (1997)
Wilson, D.: Asymptotic Properties of Nearest Neighbor Rules Using Edited Data. IEEE Trans. Systems, Man, and Cybernetics 2, 408–421 (1972)
Hu, Q.H., Xie, Z.X., Yu, D.R.: Hybrid attribute reduction based on a novel fuzzy-rough model and information granulation. Pattern Recognition 40(12), 3509–3521 (2007)
Hu, X., Cercone, N.: Data mining via discretization, generalization and rough set feature selection. Knowledge and Information Systems 1(1), 33–60 (1999)
Morsi, N.N., Yakout, M.M.: Axiomatics for fuzzy rough set. Fuzzy Sets System 100, 327–342 (1998)
Perou, C.M., Srlie, T., Eisen, M.B., et al.: Molecular portraits of human breast tumours. Nature 406, 747–752 (2000)
Slezak, D.: Degrees of conditional (in)dependence: A framework for approximate Bayesian networks and examples related to the rough set-based feature selection. Information Sciences 1789(3), 197–209 (2009)
Asuncion, A., Newman, D.J.: UCI Machine Learning Repository. University of California, School of Information and Computer Science, Irvine, CA (2007), http://www.ics.uci.edu/~mlearn/MLRepository.html
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2010 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Zhu, P., Hu, Q., Yang, Y. (2010). Weighted Nearest Neighbor Classification via Maximizing Classification Consistency. In: Szczuka, M., Kryszkiewicz, M., Ramanna, S., Jensen, R., Hu, Q. (eds) Rough Sets and Current Trends in Computing. RSCTC 2010. Lecture Notes in Computer Science(), vol 6086. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-13529-3_37
Download citation
DOI: https://doi.org/10.1007/978-3-642-13529-3_37
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-13528-6
Online ISBN: 978-3-642-13529-3
eBook Packages: Computer ScienceComputer Science (R0)