Abstract
Decision Tree algorithms cannot learn accurately with a small training set. This is because, decision tree algorithms recursively partition the data set that leaves very few instances in the lower levels of the tree. Additional domain knowledge has been shown to enhance the performance of learners. We present an algorithm named Importance Aided Decision Tree (IADT) that takes Feature Importance as an additional domain knowledge. Decision Tree algorithm always finds the most important attributes in each node. Thus, Feature Importance can be useful to Decision Tree learning. Our algorithm uses a novel approach to incorporate this feature importance score into decision tree learning. This approach makes decision trees more accurate and robust. We demonstrated theoretical and empirical performance analysis to show that IADT is superior to standard decision tree learning algorithms.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Scott, A.C., Clayton, J.E., Gibson, E.L.: A practical guide to knowledge acquisition. Addison-Wesley (1991)
Alon, N., Spencer, J.: The Probabilistic Method. Wiley, New York (1992)
Breiman, L., Freidman, J.H., Olshen, R.A., Stone, C.J.: Classification and regression trees
Esposito, F., Malerba, D., Semeraro, G.: A comparative analysis of methods for pruning decision trees. IEEE Trans. Pattern Anal. Mach. Intell. 19(5), 476–491 (1997)
Hoeffding, W.: Probability inequalities for sums of bounded random variables. J. Amer. Statist. Assoc., 13–30
Iqbal, R.A.: Empirical learning aided by weak knowledge in the form of feature importance. In: CMSP 2011. IEEE (2011)
Mitchell, T.M.: Artificial neural networks, pp. 81–126. McGraw-Hill Science/Engineering/Math. (1997)
Mitchell, T.M.: Machine Learning. McGraw-Hill (1997a)
Quinlan, J.R.: C4.5: Programs for Machine Learning. Morgan Kaufmann, San Mateo (1993)
Rokach, L., Maimon, O.: Top-down induction of decision trees classifiers - a survey. IEEE Transactions on Systems, Man, and Cybernetics, Part C 35(4), 476–487 (2005)
Vapnik, V.N.: Statistical Learning Theory. Wiley, New York (1998)
Zhang, L., Wang, Z.: Ontology-based clustering algorithm with feature weights. Journal of Computational Information Systems 6(9) (2010)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2012 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Iqbal, M.R.A., Rahaman, M.S., Nabil, S.I. (2012). Construction of Decision Trees by Using Feature Importance Value for Improved Learning Performance. In: Huang, T., Zeng, Z., Li, C., Leung, C.S. (eds) Neural Information Processing. ICONIP 2012. Lecture Notes in Computer Science, vol 7664. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-34481-7_30
Download citation
DOI: https://doi.org/10.1007/978-3-642-34481-7_30
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-34480-0
Online ISBN: 978-3-642-34481-7
eBook Packages: Computer ScienceComputer Science (R0)