Skip to main content

Selection of a relevant feature subset for induction tasks

  • Communications
  • Conference paper
  • First Online:
Foundations of Intelligent Systems (ISMIS 1999)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 1609))

Included in the following conference series:

Abstract

The representation of problems dealt with by machine learning systems use many features, only a few of which may be related to concept designing. Feature selection is the problem of choosing an ideally small subset of necessary features that are sufficient to describe the target concept. It is important both to speed up learning and to improve concept quality. A huge amount of work has been done to select from input data, a subset of the most relevant features. In this paper, a new algorithm of feature pre-processing for induction methods, namely induction of decision trees, and applied to symbolic objects is suggested. It selects a subset of the more relevant features, taking into account feature interaction. It is based both on the DPGoal and ODPGoal of considered variables (features). It is evaluated using three benchmark artificial domains. Then it is compared with Relief [1].

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Kira, K., Rendell, L. A.: The Feature Selection Problem: Traditional Methods and a New Algorithm. Proc. AAAI-92, San Jose (1992)

    Google Scholar 

  2. Dash, M., Liu, H.: Feature Selection for Classification. Intelligent Data Analysis, Vol. 1, no. 3 (1997)

    Google Scholar 

  3. Bratko, I., Cestnik, B., Kononenko, I.: Attribute-based learning. AI Communications, Vol. 9 (1996), 27–32.

    Google Scholar 

  4. Breiman, L., Friedman, J. H., Olshen, R. A., Stone, C. J.: Classification and Regression Trees. Wadsworth, (1984)

    Google Scholar 

  5. Buntine, W., Niblett, T.: A Further Comparison of Splitting Rules for Decision Tree Induction. Machine Learning, Vol. 8 (1992), 75–85

    Google Scholar 

  6. White, A. P., Liu, W. Z.: Bias in Information-Based Measures in Decision Tree Induction. Machine Learning, Vol. 15 (1994), 321–329.

    MATH  Google Scholar 

  7. Mingers, J.: An Empirical Comparison of Selection Measures for Decision Tree Induction. Machine Learning, 3 (1989), 319–342.

    Google Scholar 

  8. Liu, W. Z., White, A. P.: The Importance of Attribute Selection Measures in Decision Tree Induction. Machine Learning, Vol. 15 (1994), 25–41

    Google Scholar 

  9. Quinlan, J. R., C4.5: programs for machine learning, Morgan Kaufmann (1993)

    Google Scholar 

  10. Langley, P., Sage, S.: scaling to Domains with Irrelevant features. Computational learning theory and natural learning systems, Vol. 4. MA: MIT Press, Cambridge (1997), 17–29

    Google Scholar 

  11. Kira, K., Rendell, L. A.: A Practical Approach to Feature Selection. Machine Learning (1992), 249–255

    Google Scholar 

  12. Almuallim, H., Dietterich, T. G.: Learning With Many Irrelevant Features. Ninth National Conference on Artificial Intelligence (1991), 547–552

    Google Scholar 

  13. Dougherty, J., Kohavi, R., Sahami, M.: Supervised and Unsupervised Discretization of Continuous Features. Proc. of the 12th ICML, San Francisco (1994), 194–202

    Google Scholar 

  14. Brito, P., Diday, E.: Pyramidal representation of symbolic objects. In: Schalder, M. & Gaul, W. (eds.): Knowledge, Data and Computer-Assisted Decisions Springer-Verlag, Berlin (1990), 3–16.

    Google Scholar 

  15. Kohavi, R., John, G. H.: Wrappers for feature subset selection. Artificial Intelligence (1997), 273–324.

    Google Scholar 

  16. Ziani, D., Khalil, Z., Vignes, R.: Recherche de sous-ensembles minimaux de variables à partir d’objets symboliques. Proc. 5 èmes journées “Symbolique-Numérique”, IPMU-94, Paris (1994), 794–799.

    Google Scholar 

  17. Smyth, P., Goodman, R.M., Higgins, C.: A hybrid rule-based Bayesian classifier. Proc. of ECAI, Stockholm (1990), 610–615

    Google Scholar 

  18. Blum, A. L. LLangley, P.: Selection of relevant features and examples in machine learning. Artificial Intelligence (1997), 245–271

    Google Scholar 

  19. Kononenko, I.: Estimating Attributes: Analysis and Extensions of RELIEF. ECML-94, Catania, Italy (1994), 171–182

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Zbigniew W. Raś Andrzej Skowron

Rights and permissions

Reprints and permissions

Copyright information

© 1999 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Michaut, D., Baptiste, P. (1999). Selection of a relevant feature subset for induction tasks. In: Raś, Z.W., Skowron, A. (eds) Foundations of Intelligent Systems. ISMIS 1999. Lecture Notes in Computer Science, vol 1609. Springer, Berlin, Heidelberg . https://doi.org/10.1007/BFb0095134

Download citation

  • DOI: https://doi.org/10.1007/BFb0095134

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-65965-5

  • Online ISBN: 978-3-540-48828-6

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics