Skip to main content

Combining Feature Selection with Feature Weighting for k-NN Classifier

  • Conference paper
  • First Online:
Intelligent Data Engineering and Automated Learning — IDEAL 2002 (IDEAL 2002)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 2412))

Abstract

The k-nearest neighbor (k-NN) classification is a simple and effective classification approach. However, it suffers from over-sensitivity problem due to irrelevant and noisy features. In this paper, we propose an algorithm to improve the effectiveness of k-NN by combining these two approaches. Specifically, we select all relevant features firstly, and then assign a weight to each one. Experimental results show that our algorithm achieves the highest accuracy or near to the highest accuracy on all test datasets. It also achieves higher generalization accuracy compared with the well-known algorithms IB1-4 and C4.5.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. D.W. Aha, D. Kibler & M.K. Albert, “Instance-Based Learning Algorithms”, Machine Learning, Vol. 6, 1991,pp. 37–66.

    Google Scholar 

  2. D.W. Aha, “Tolerating Noisy, Irrelevant and Novel Attributes in Instance-Based Learning Algorithm”, International Journal of Man-Machine Studies, Vol. 36, No. 1, 1992, pp. 267–287.

    Article  Google Scholar 

  3. D.W. Aha & R.L. Bankert, “Feature Selection for Case-Based Classification of Cloud Types: An Empirical Comparison”, In D.W. Aha (Ed.) Case-Based Reasoning: Papers from the 1994 Worksho ((TR WS-94-10), Menlo Park, CA: AAAI press.

    Google Scholar 

  4. D.W. Aha, “Feature Weighting for Lazy Learning Algorithms”, In H. Liu & H. Motoda (Eds.) Feature Extraction, Construction and Selection: A Data Mining Perspective, Norwell MA: Kluwer, 1998.

    Google Scholar 

  5. N.F. Ayan, “Using Information Gain as Feature Weight”, 8th Turkish Symposium on Artificial Intelligence and Neural Networks (TAINN’99), Istanbul, Turkey, June 1999.

    Google Scholar 

  6. Y. Bao, X. Du & N. Ishii, “Using Rough Sets and Class Mutual Information for Feature Selection”, 6th International Conference on Soft Computing, IIZUKA2000, pp. 452–458.

    Google Scholar 

  7. R. Battiti, “Using Mutual Information for Selection Features in Supervised Neural Net Learning,” IEEE Trans. On Neural Networks, 5(4), 1994, pp. 537–550.

    Article  Google Scholar 

  8. C. Cardie, “A Cognitive Bias Approach to Feature Selection and Weighting for Case-Based Learners”, Machine Learning, Vol. 41, 2000, pp. 85–116.

    Article  Google Scholar 

  9. T.M. Cover & P.E. Hart, “Nearest Neighbor Pattern Classification”, IEEE Transactions on Information Theory, Vol. 13, No. 1, 1967, pp. 21–27.

    Article  MATH  Google Scholar 

  10. T.M. Cover & J.A. Thomas, Elements of Information Theory, 1991.

    Google Scholar 

  11. R. Kohavi, P. Langley & Y. Yun, “The Utility of Feature Weighting in Nearest Neighbor Algorithms”, ECML-97 (poster).

    Google Scholar 

  12. K.C. Lee, “A Technique of Dynamic Feature Selection Using the Feature Group Mutual Information”, Third Pacific Asia Conference, PAKDD-99, LNAI 1574, 1999, pp. 138–142.

    Google Scholar 

  13. C. J. Merz & P.M. Murphy, 1998, UCI Repository of Machine Learning Databases, Irvine, CA: University of California Irvine, Department of Information and Computer Science, Internet: http://www.ics.uci.edu/ mlearn/MLRepository.html.

    Google Scholar 

  14. Z. Pawlak, ROUGH SETS (Theoretical Aspects of Reasoning about Data), Kluwer (1991).

    Google Scholar 

  15. J.R. Quinlan, “Induction of Decision Trees”. Machine Learning 1(1), 1986, pp. 81–106.

    Google Scholar 

  16. J.R. Quinlan, C4.5: Programs for Machine Learning, Morgan Kaufmann, 1993.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2002 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Bao, Y., Du, X., Ishii, N. (2002). Combining Feature Selection with Feature Weighting for k-NN Classifier. In: Yin, H., Allinson, N., Freeman, R., Keane, J., Hubbard, S. (eds) Intelligent Data Engineering and Automated Learning — IDEAL 2002. IDEAL 2002. Lecture Notes in Computer Science, vol 2412. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-45675-9_69

Download citation

  • DOI: https://doi.org/10.1007/3-540-45675-9_69

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-44025-3

  • Online ISBN: 978-3-540-45675-9

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics