Abstract
k-nearest neighbor method (kNN) is a very useful and easy-implementing method for real applications. The query point is estimated by its k nearest neighbors. However, this kind of prediction simply uses the label information of its neighbors without considering their space distributions. This paper proposes a novel kNN method in which the centroids instead of the neighbors themselves are employed. The centroids can reflect not only the label information but also the distribution information of its neighbors. In order to evaluate the proposed method, Euclidean distance and Mahalanobis distance is used in our experiments. Moreover, traditional kNN is also implemented to provide a comparison with the proposed method. The empirical results suggest that the propose method is more robust and effective.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Yang, L.: Distance Metric Learning: A Comprehensive Survey. Department of Computer Science and Engineering. Michigan State University (2006)
Cover, T., Hart, P.: Nearest Neighbor Pattern Classification. IEEE Transaction on Information Theory, 21–27 (1967)
Duda, R.O., Hart, P.E.: Pattern Classification and Scene Analysis. Wiley, New York (1973)
Achtert, E., Kriegel, H.P., Kröger, P., Renz, M., Züfle, A.: Reverse k-Nearest Neighbor Search in Dynamic and Deneral Metric Databases. In: Proceedings of the 12th International Conference on Extending Database Technology, pp. 886–897 (2009)
Sun, S., Huang, R.: An Adaptive k-nearest Neighbor Algorithm. In: Proceedings of the 7th International Conference on Fuzzy Systems and Knowledge Discovery, pp. 91–94 (2010)
Weinberger, K.Q., Saul, L.K.: Distance Metric Learning for Large Margin Nearest Neighbor Classification. Journal of Machine Learning Research 10, 207–244 (2009)
Jin, C., Guo, W.: Efficiently Monitoring Nearest Neighbors to a Moving Object. In: Proceedings of International Conference on Advanced Data Mining and Applications, pp. 239–251 (2007)
Sun, S.: Local Within-class Accuracies for Weighting Individual Outputs in Multiple Classifier Systems. Pattern Recognition Letters 31(2), 119–124 (2010)
Hastie, T., Tibshirani, R.: Discriminant Adaptive Nearest Neighbor Classification. IEEE Pattern Analysis and Machine Intelligence 18(6) (1996)
Friedman, J.: Flexible Metric Nearest Neighbor Classification. Stanford University Statistics Department, Tech. Rep. (1994)
Duda, R.O., Hart, P.E., Stork, D.G.: Pattern Classification, 2nd edn. John Wiley and Sons, New York (2001)
Ng, A.Y., Jordan, M.I., Weiss, Y.: On Spectral Clustering: Analysis and An Algorithm. In: Advances in Neural Information Processing Systems (2001)
MacKay, D.J.C.: Information Theory, Inference and Learning Algorithms. Cambridge University Press, Cambridge (2003)
Kanungo, T., Mount, D.M., Netanyahu, N.S., Piatko, C.D., Silverman, R., Wu, A.Y.: An Efficient K-means Clustering Algorithm: Analysis and implementation. IEEE Transaction on Pattern Analysis and Machine Intelligence 24, 881–892 (2002)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2010 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Zhang, Q., Sun, S. (2010). A Centroid k-Nearest Neighbor Method. In: Cao, L., Feng, Y., Zhong, J. (eds) Advanced Data Mining and Applications. ADMA 2010. Lecture Notes in Computer Science(), vol 6440. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-17316-5_27
Download citation
DOI: https://doi.org/10.1007/978-3-642-17316-5_27
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-17315-8
Online ISBN: 978-3-642-17316-5
eBook Packages: Computer ScienceComputer Science (R0)