Elsevier

Information Sciences

Volume 248, 1 November 2013, Pages 130-150
Information Sciences

Nested structure in parameterized rough reduction

https://doi.org/10.1016/j.ins.2013.05.039Get rights and content

Abstract

In this paper, by strict mathematical reasoning, we discover the relationship between the parameters and the reducts in parameterized rough reduction. This relationship, named the nested reduction, shows that the reducts act as a nested structure with the monotonically increasing parameter. We present a systematic theoretical framework that provides some basic principles for constructing the nested structure in parameterized rough reduction. Some specific parameterized rough set models in which the nested reduction can be constructed are pointed out by strict mathematical reasoning. Based on the nested reduction, we design several quick algorithms to find a different reduct when one reduct is already given. Here ‘different’ refers to the reducts obtained on the different parameters. All these algorithms are helpful for quickly finding a proper reduct in the parameterized rough set models. The numerical experiments demonstrate the feasibility and the effectiveness of the nested reduction approach.

Introduction

Rough set (RS) theory, which was proposed by Pawlak [24], [25], is a mathematical tool to handle uncertainty of indiscernibility. RS is effective in many real applications such as artificial intelligence, data mining and pattern recognition. However, it is limited by its basic definition ‘equivalence relation’. As a result, many generalizations, such as fuzzy rough sets [7], [8], [22], [59], [61], cover rough sets [48], [49], [56], Bayesian rough sets [18], [32], [36], [54] and probabilistic rough sets [42], [43], [51], [55], have been proposed. These generalizations make rough sets feasible to handle many types of practical problems such as the problems with real values, the problems with missing values and the problems with random uncertainty.

One of the important applications of rough set theory is attribute reduction [11], [14], [34], [37], [45], [47], [62], [63]. In recent years, researchers, motivated by a desire to acquire reduction robustly, have proposed many methods to mining valuable and less-sensitive attributes by incorporating parameters into rough set theory [15], [18], [21], [57], [58], [60]. These methods are called parameterized rough reduction. Roughly speaking, parameterized rough reduction is split into two types: reduction on parameterized rough models and parameterized reduction on rough sets. The parameterized rough models share a common characteristic by introducing parameters into the rough approximation operators, whereas parameterized reduction deletes redundant attributes by introducing thresholds into the process of attribute reduction.

Many parameterized rough models are presently being proposed and studied intensively, they are Variable Precision Rough Set Models [2], [3], [50], [53], Robust Fuzzy Rough Set Models [1], [9], [10], [26], [45], Probabilistic Rough Set Models [42], [43], [51], Decision Theoretic Rough Set Models [17], [39], [40], and Bayesian Rough Set Models [18], [32]. The Variable Precision Rough Set Models treat the required parameters as a primitive notion [50]. The interpretation and the process of determination of parameters are based on rather intuitive arguments and left to empirical studies [2], [3], [16], with the lack of theoretical and systematic studies and justifications on the choices of the threshold parameters. K-Nearest Neighbor Fuzzy Rough Sets provide some methods to determine the required parameters [10], whereas the applications of these models focus mainly on making classification predications. Unlike the aforementioned models, the Bayesian Rough Set Models [18], [32] attempt to provide an alternative interpretation of the required parameters. The models are based on Bayes’ Rule that expresses the change from the a priori probability to the a posteriori probability [40], [41]. In these models, the required parameters can be expressed in terms of probabilities. In addition to variable precision analysis and Bayesian analysis, the parameterized approaches have been applied to the theory of rough sets in some other forms, such as decision-theoretic analysis [39], [40] and probabilistic analysis [42], [43], [51]. The probabilistic rough set models and decision-theoretic rough set models provide a unified and comprehensive framework so that many types of parameterized models can be integrated into a whole [38]. These models provide a systematic method for determining the parameters by using more concrete notions of costs and risks.

Instead of introducing parameters into the approximation operators, parameterized reduction on rough sets introduces parameters into the process of attribute reduction. By relaxing the criteria of attribute reduction, some parameters are introduced in the processing (or notions) of attribute reduction, such as the method of fuzzy reductions (Fuzzy-RED) [5], [6] and approximate reductions (Approximate-RED) [28], [29], [30]. The common characteristic of these methods is that the parameters are put into the process of attribute reduction. The main difference between these models is that the distinct measures of discernibility power are used in the process of attribute reduction, such as dependency function, information entropy and monotonic measure. The determination of these parameters is usually adopted in an intuitive way.

Whether reduction on parameterized models or parameterized reduction on rough sets is applied, a difficulty with many existing studies on parameterized rough reductions is that no interpretation or procedure for calculating the required threshold exists. In real applications, the setting of parameters affects the results of the reduction significantly, and different thresholds may lead to different reduction results. Some researchers realize the significance of the parameters, and they have paid attention to the parameter analysis and selection. Intuitively, the parameters are given by the experience of an expert in most of the parameterized rough reductions [11], [12], [52]. Some parameterized rough reduction approaches, based on the assumption that the threshold is in an interval that was determined by the desired level of classification performance, proposed methods for finding an appropriate reduct [2], [3]. Other researchers have tried to find a reasonable threshold based on an assessment of the minimum acceptable upper bound of the misclassification error [33]. Essentially, these approaches utilize the principle of the ‘extent of classification correctness’ to determine an appropriate threshold value under the premise that changing the number of attributes has no effect on the classification results [13]. However, such approaches need to set the classification error beforehand. Instead of setting the parameters intuitively, a systematic method for computing parameters in probabilistic rough sets was provided. Yao et al. introduced risk to probabilistic rough sets and proposed decision theoretic rough sets with Bayesian decision procedures [38], [41].These studies have explicitly noted the parameter effect on the values of approximation operators. Although the aforementioned methods adopt and design different methods of determining and interpreting the parameters, these methods share a common characteristic that all of them seek one optimal or suboptimal value to be the required parameter. These methods omit one possible situation that often occurs in real applications: the required parameter changes frequently. That is, the required parameters are not fixed, and these parameters often change with time or with other conditions in real application.

Some researchers have realized the importance of the parameters and have described various methods for setting parameters. Unfortunately, many researchers are still unaware of the connection between reduction and parameters. The existing methods do not indicate how the parameters affect the performance of reduction. Nobody has mentioned the approaches that can determine the proper reduct based on the given reducts. Discovering the inner relationship between the parameters and the reducts in parameterized rough reduction approaches is now a promising and necessary area of research. In this paper, we propose such a method to quickly find reducts on different parameters in parameterized rough reduction by explicitly showing the connections between the reducts and the parameters.

The remainder of this paper is organized as follows. Section 2 gives some preliminaries, such as rough sets and a parameterized rough reduction. After reviewing many approaches to parameterized rough reduction, we discover the inner connection between the reducts and the parameters in Section 3. By strict mathematical reasoning, we describe the connection, which is called the nested structure. A systematic approach to identify the nested structure is proposed based on a strict mathematical foundation. By using the nested reduction structure, Section 4 proposes some algorithms to find the required reducts quickly. By using numerical experiments, Section 5 clearly demonstrates the effectiveness of these algorithms. Finally, a conclusion is drawn in Section 6.

Section snippets

Parameterized rough reduction

In this section, we present some methods to address parameterized rough reduction, that is, reduction in parameterized rough models and parameterized reduction on rough sets.

The nested structure in parameterized rough reduction

In this section, the nested reduction, which defines the inner relationship between the parameters and the reducts in parameterized rough reduction, is described. First, some definitions of the nested reduction are proposed that describe the nested structure existing in parameterized rough reduction. Then, some theorems are presented. These theorems provide principles that are needed to construct the nested reduction structure. Furthermore, these theorems show that not all parameterized rough

Some quick reduction algorithms

In this section, by using the nested structure of reduction, we design some algorithms to find a different reduct quickly when a reduct is given. To fasten the proposed algorithms, we improve them by finding the different reducts with no repetition. The existing algorithms to find the suboptimal reduct are also given.

Numerical experiments

In this section, several datasets with different object numbers, decision classes, condition attributes and data distributions (see Table 3) are selected to demonstrate the performance of the algorithms designed based on the nested reduction.

As mentioned previously in Section 3, there are many parameterized models suited for the construction of the nested or weakly nested reductions. Here, FVPRS, VPBRS and fuzzy-RED, which determine reducts by using three different types of attribute reduction

Conclusions

This paper studies the structure of the reducts on the different parameters in parameterized rough sets. The authors discover that the reducts on different parameters are often nested or the same. In this paper, it is called the nested structure. Based on this interesting discovery, some theorems and algorithms have been proposed. The theorems show, by using strict mathematical reasoning, that the nested structure exactly exists in parameterized rough reduction. Using several theorems and

Acknowledgments

This research is supported by the National Basic Research Program of China (973 Program) (2012CB316205), the National Science Foundation of China (61202114, 61272137, 61070056, 61033010, 61170040 and 60903089). This research is also supported by the Fundamental Research Funds for the Central Universities, and the Research Funds of Renmin University of China (12XNLF07). We also acknowledge the anonymous reviewers for their valuable comments and suggestions.

References (63)

  • D. Slezak et al.

    The investigation of the Bayesian rough set model

    Int. J. Approx. Reason.

    (2005)
  • C.T. Su et al.

    Precision parameter in the variable precision rough sets model: an application

    OMEGA – Int. J. Manag. Sci.

    (2006)
  • W.Z. Wu et al.

    Knowledge reduction in random information systems via Dempster–Shafer theory of evidence

    Inf. Sci.

    (2005)
  • Y.Y. Yao

    Probabilistic rough set approximations

    Int. J. Approx. Reason.

    (2008)
  • Y.Y. Yao et al.

    Attribute reduction in decision-theoretic rough set models

    Inf. Sci.

    (2008)
  • Y.Y. Yao

    Three-way decisions with probabilistic rough sets

    Inf. Sci.

    (2010)
  • Y.Y. Yao

    The superiority of three-way decisions in probabilistic rough set models

    Inf. Sci.

    (2011)
  • W. Zhu

    Topological approaches to covering rough sets

    Inf. Sci.

    (2007)
  • W. Ziarko

    Variable precision rough set model

    J. Comput. Syst. Sci.

    (1993)
  • W. Ziarko

    Probabilistic approach to rough sets

    Int. J. Approx. Reason.

    (2008)
  • D.G. Chen et al.

    Parameterized attribute reduction with Gaussian kernel based fuzzy rough sets

    Inf. Sci.

    (2011)
  • X.Y. Zhang et al.

    Comparative study of variable precision rough set model and graded rough set model

    Int. J. Approx. Reason.

    (2012)
  • H.Y. Zhang et al.

    Bayesian rough set model: a further investigation

    Int. J. Approx. Reason.

    (2012)
  • W. Zhu et al.

    The fourth type of covering-based rough sets

    Inf. Sci.

    (2012)
  • H.Y. Zhang et al.

    Variable-precision-dominance-based rough set approach to interval-valued information systems

    Inf. Sci.

    (2013)
  • W. Wei et al.

    A comparative study of rough sets for hybrid data

    Inf. Sci.

    (2012)
  • S.M. Vieira et al.

    Fuzzy criteria for feature selection

    Fuzzy Sets Syst.

    (2012)
  • N. Verbiest et al.

    FRPS: a fuzzy rough prototype selection method

    Pattern Recog.

    (2013)
  • S. Mitra et al.

    Feature selection using structural similarity

    Inf. Sci.

    (2012)
  • S. An et al.

    Soft minimum-enclosing-ball based robust fuzzy rough sets

    Fundam. Inform.

    (2012)
  • M. Beynon, An investigating of beta-reduct selection within the variable precision rough sets model, in: The 2nd Int....
  • Cited by (26)

    • PARA: A positive-region based attribute reduction accelerator

      2019, Information Sciences
      Citation Excerpt :

      For another example, mRMR (maximum relevance minimum redundancy) [30] is a mutual information-based filter feature selection method for finding a set of relevant and complementary features. What is more, rough/fuzzy rough reduction is a useful attribute reduction method [1,3,4,5,16,18,28,37,38,39,43,48,49], whose characteristics are human understanding and non-need extra expert knowledge. Fuzzy rough sets (FRS) are the generalization of rough sets in fuzzy set framework, which assume that instances characterized by the same information are indiscernible (similar) in the view of the available information about them (with every instance in the Universe we associate some information) [9,22,28,29,45].

    • Uncertainty learning of rough set-based prediction under a holistic framework

      2018, Information Sciences
      Citation Excerpt :

      Uncertainty learning is the key problem of rough set theory, which was originally proposed by Pawlak [33] in 1982 as a mathematical approach for data analysis and knowledge discovery. Rough set has been successfully applied to many real-world problems in machine learning [51], data mining [10,49], pattern recognition, decision analysis [1,20,50], expert systems [21,53], intelligent control [4,42] and so on. Rough sets extract decision rules from decision systems by employing attribute reduction [35] and feature selection [13,30,31] as a processing step, because decision system contains the repeating and inconsistent samples.

    • Parallel attribute reduction in dominance-based neighborhood rough set

      2016, Information Sciences
      Citation Excerpt :

      The degree of preference between numerical and categorical data is taken into consideration. Attribute reduction is one of the most important applications of RST, and it has been studied extensively [9,22,24,31,33,40,43,45,47,51,53,59]. There are many novel attribute reduction methods which have considered decision regions [30,50], dynamic properties of the information system [23,54,55], combining with other methods [8,41], dealing with hybrid data [19,56], and giving a final unification of interclass reductions, intraclass reductions and constructs across CRSA and DRSA [44].

    • A novel attribute reduction approach for multi-label data based on rough set theory

      2016, Information Sciences
      Citation Excerpt :

      Feature selection in rough set theory is also called attribute reduction; it aims to remove unnecessary attributes while retaining the discernibility of objects under the original attributes. In the past few years, many types of attribute reduction approaches have been proposed according to various criteria [3,4,8,11,17,22,33,42,44,47,49,52,53]. For convenience, some of these techniques are briefly reviewed here.

    • Rough approximation of a preference relation by multi-decision dominance for a multi-agent conflict analysis problem

      2015, Information Sciences
      Citation Excerpt :

      Because all decision-making problems have multiple alternatives and criteria, an increase in the number of alternatives and criteria makes it difficult to make decisions. Therefore, there has been extensive research on how to solve complicated decision making problems [8,18,35,39,40,49,50,54,56,57,63,73,74]. In general, the problem of multi-decision with preference (or multiple attributes and multiple decisions with preference (MA&MD)) is the most frequently considered decision problem in decision analysis [10,11,16,17,15,38,41,45,60,61].

    View all citing articles on Scopus
    View full text