skip to main content
10.1145/775047.775113acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
Article

Extracting decision trees from trained neural networks

Published:23 July 2002Publication History

ABSTRACT

Neural Networks are successful in acquiring hidden knowledge in datasets. Their biggest weakness is that the knowledge they acquire is represented in a form not understandable to humans. Researchers tried to address this problem by extracting rules from trained Neural Networks. Most of the proposed rule extraction methods required specialized type of Neural Networks; some required binary inputs and some were computationally expensive. Craven proposed extracting MofN type Decision Trees from Neural Networks. We believe MofN type Decision Trees are only good for MofN type problems and trees created for regular high dimensional real world problems may be very complex. In this paper, we introduced a new method for extracting regular C4.5 like Decision Trees from trained Neural Networks. We showed that the new method (DecText) is effective in extracting high fidelity trees from trained networks. We also introduced a new discretization technique to make DecText be able to handle continuous features and a new pruning technique for finding simplest tree with the highest fidelity.

References

  1. L. Atlas, R. Cole, J. Connor, M. El-Sharkawi, R. J. Marks, V. Muthusamy, and E. Barnard. Performance comparisons between backpropagation networks and classification trees on three real-world applications. In Advances in Neural Information Processing Systems, volume 2, pages 622--629, 1990. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. O. Boz. Converting A Trained Neural Network To A Decision Tree DecText - Decision Tree Etxractor. PhD thesis, Computer Science and Engineering, Lehigh University, 2000. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. W. Buntine and T. Niblett. A further comparison of splitting rules for decision-tree induction. Machine Learning, 8:75, 1992. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. M. W. Craven. Extracting Comprehensible Models from Trained Neural Networks. PhD thesis, Department of Computer Sciences, University of Wisconsin-Madison, 1996. (Also appears as UW Technical Report CS-TR-96-1326). Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. M. W. Craven and J. W. Shavlik. Extracting tree-structured representations of trained networks. In Advances in Neural Information Processing Systems, volume 8, pages 24--30, Denver, CO, 1996. MIT Press.Google ScholarGoogle Scholar
  6. D. Fisher and K. McKusick. An emprical comparison of id3 and back propagation. In Proceedings of the eleventh international Joint Conference on Artificial Intelligence, pages 788--793, Detroit, 1989. Morgan Kaufmann.Google ScholarGoogle Scholar
  7. P. Murphy and M. Pazzani. Id2-of-3: Constructive induction of n-of-m concepts for discriminators in decision trees. In Proceedings of the Eighth International Machine learning Workshop, pages 183--187, Evanston, IL, 1991. Morgan Kaufmann.Google ScholarGoogle ScholarCross RefCross Ref
  8. J. W. Shavlik, R. J. Mooney, and G. G. Towell. Symbolic and neural learning algorithms: An experimental comparison. Machine Learning, 6:111--143, 1991. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. B. W. Silverman. Density Estimation for Statistics and Data Analysis. Chapman & Hall, London, 1986.Google ScholarGoogle ScholarCross RefCross Ref
  10. S. M. Weiss and I. Kapouleas. An empirical comparison of pattern recognition, neural nets, and machine learning classification methods. In J. W. Shavlik and T. G. Dietterich, editors, Readings in Machine Learning. Morgan Kaufman, San Mateo, CA, 1990.Google ScholarGoogle Scholar

Index Terms

  1. Extracting decision trees from trained neural networks

                Recommendations

                Comments

                Login options

                Check if you have access through your login credentials or your institution to get full access on this article.

                Sign in
                • Published in

                  cover image ACM Conferences
                  KDD '02: Proceedings of the eighth ACM SIGKDD international conference on Knowledge discovery and data mining
                  July 2002
                  719 pages
                  ISBN:158113567X
                  DOI:10.1145/775047

                  Copyright © 2002 ACM

                  Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

                  Publisher

                  Association for Computing Machinery

                  New York, NY, United States

                  Publication History

                  • Published: 23 July 2002

                  Permissions

                  Request permissions about this article.

                  Request Permissions

                  Check for updates

                  Qualifiers

                  • Article

                  Acceptance Rates

                  KDD '02 Paper Acceptance Rate44of307submissions,14%Overall Acceptance Rate1,133of8,635submissions,13%

                  Upcoming Conference

                PDF Format

                View or Download as a PDF file.

                PDF

                eReader

                View online with eReader.

                eReader