Elsevier

Fuzzy Sets and Systems

Volume 99, Issue 3, 1 November 1998, Pages 283-290
Fuzzy Sets and Systems

On the handling of fuzziness for continuous-valued attributes in decision tree generation

https://doi.org/10.1016/S0165-0114(97)00030-4Get rights and content

Abstract

In this paper, fuzziness existing in the process of generating decision trees by discretizing continuous-valued attributes is considered. In a sense a better way to express this fuzziness via fuzzy numbers is presented using possibility theory. The fact that selection of membership functions in a class of symmetric distributions does not influence the decision tree generation is proved. The validity of using the tree to classify future examples is explained. On the basis of likelihood possibility maximization, the existing algorithm is revised. The revised algorithm leads to more reasonable and more natural decision trees.

References (12)

  • W. Nather

    On possibilistic inference

    Fuzzy Sets and Systems

    (1990)
  • W. Zhenyuan et al.

    Fuzzy linear regression analysis of fuzzy-valued variables

    Fuzzy Sets and Systems

    (1990)
  • J. Cheng et al.

    Improved decision trees: a generalized version of ID3

  • P. Clark et al.

    The CN2 induction algorithm

    Machine Learning

    (1989)
  • D. Dubois et al.

    Possibility Theory

    (1988)
  • U.M. Fayyad et al.

    A machine learning algorithm (GID3) for automated knowledge acquisition: improvements and extensions

    (1991)
There are more references available in the full text version of this article.

Cited by (54)

  • Fuzzy nonlinear regression analysis using a random weight network

    2016, Information Sciences
    Citation Excerpt :

    Second, we will theoretically analyze the influence of random initialization on the approximation power of FNRRWN. Third, we will develop the fuzzy RWNs based on granular computing paradigm [28] and maximum uncertainty [36,37,39,41]. Fourth, we will find some practical applications for FNRRWN, e.g., image processing [24,44,55], sentiment analysis for Chinese text [9–11], and protein-protein interaction prediction [46,47].

  • Learning ELM-Tree from big data based on uncertainty reduction

    2015, Fuzzy Sets and Systems
    Citation Excerpt :

    Generally, the scale of a tree is affected by two factors, i.e., the degree of over-partitioning in the induction process, and the selection of heuristic measure for splitting nodes. Information entropy [18,22] and ambiguity [29,34], which respectively reflect the impurity of classes and the uncertainty of the split, are the two most widely used heuristics for splitting nodes during the tree growth. The performances of these two heuristics may differ a lot, however, both of them are uncertainty reduction based methods [31,32] that can be applied to the parallelization of sequential DT.

  • Dynamic optimisation based fuzzy association rule mining method

    2019, International Journal of Machine Learning and Cybernetics
  • Certain types of soft coverings based rough sets with applications

    2019, International Journal of Machine Learning and Cybernetics
View all citing articles on Scopus
View full text