Skip to main content

Estimating feature discriminant power in decision tree classifiers

  • Posters
  • Conference paper
  • First Online:
  • 128 Accesses

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 970))

Abstract

Feature Selection is an important phase in pattern recognition system design. Even though there are well established algorithms that are generally applicable, the requirement of using certain type of criteria for some practical problems makes most of the resulting methods highly inefficient. In this work, a method is proposed to rank a given set of features in the particular case of Decision Tree classifiers, using the same information generated while constructing the tree. The preliminary results obtained with both synthetic and real data confirm that the performance is comparable to that of sequential methods with much less computation.

This work has been partially supported by project P1A94-23 Fundació Caixa Castelló, and project GV-2110/94 Conselleria d'Educació i Ciència, Generalitat Valenciana.

This is a preview of subscription content, log in via an institution.

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Boswell, R.; Manual for NewID, version 5.1, The Turing Institute, Ref. TI/P2154/RAB/4/2.4.

    Google Scholar 

  2. Breiman L. et. al., Classification and Regression Trees, Chapman & Hall, 1984

    Google Scholar 

  3. Brodatz, Textures: A Photographic Album for Artists and Designers, Dover Publications, New York, 1966

    Google Scholar 

  4. Devijver, P.A. and Kittler, J. Pattern Recognition: a Statistical Approach, Prentice-Hall International, 1982

    Google Scholar 

  5. Guo, H. and Genlfand, S.B. “Classification Trees with Neural Network Feature Extraction”, IEEE Trans. on Neural Networks, Vol. 3, No. 6, pp. 923–933, 1992.

    Google Scholar 

  6. Kittler, J. “Feature Selection and Extraction”, Handbook of Pattern Recognition and Image Processing, 1986

    Google Scholar 

  7. Murthy, S.K.; Kasif, S. and Salzberg, S.; “A System for Induction of Oblique Decision Trees”, Journal of Artificial Intelligence Research, 2, 1994, pp. 1–32.

    Google Scholar 

  8. Narendra, P.M. and Fukunaga, K. “A Branch and Bound Algorithm for Feature Subset Selection”, IEEE Trans. Comput., vol. C-26, pp 917–922, Sept. 1977

    Google Scholar 

  9. Pla, F. Estudios de Técnicas de Análisis de Imagen en un Sistema de Visión para la Recolección Robotizada de Cítricos, (in Spanish) Ph.D. Thesis, Universitat de València, 1993.

    Google Scholar 

  10. Pudil, P.; Ferri, F., Novoviçová, J and Kittler J. “Floating Search Methods for Feature Selection with Nonmonotonic Criterion Functions”, in Proc. of the 12th Intl. Conf. on Pattern Recognition, Jerusalem, 1994.

    Google Scholar 

  11. Quinlan, J.R. “Simplifying decision trees”, International Journal of Man-Machine Studies 27, pp. 221–234, 1987.

    Google Scholar 

  12. Siedlecki, W. and Sklansky, J. “On Automatic Feature Selection”, International Journal of Pattern Recognition and Artificial Intelligence, Vol. 2, pp. 197–220, 1988.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Václav Hlaváč Radim Šára

Rights and permissions

Reprints and permissions

Copyright information

© 1995 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Gracia, I., Pla, F., Ferri, F.J., García, P. (1995). Estimating feature discriminant power in decision tree classifiers. In: Hlaváč, V., Šára, R. (eds) Computer Analysis of Images and Patterns. CAIP 1995. Lecture Notes in Computer Science, vol 970. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-60268-2_353

Download citation

  • DOI: https://doi.org/10.1007/3-540-60268-2_353

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-60268-2

  • Online ISBN: 978-3-540-44781-8

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics