Skip to main content

Attribute selection for incomplete decision systems by maximizing correlation and independence with mutual granularity

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

Rough set theory has been widely used in attribute selection. However, there are few researchers who have explored the relationship between attributes from the perspective of knowledge granularity. Additionally, existing attribute selection methods are mostly tailored for complete decision systems and are not applicable to incomplete ones. In light of the aforementioned challenge, this paper primarily focuses on addressing the issue of attribute selection for incomplete decision systems by utilizing the correlation among attributes formed through knowledge granularity. Firstly, the concept of mutual granularity is defined by introducing discernment granularity and conditional discernment granularity into incomplete decision systems. Secondly, an attribute selection algorithm based on mutual granularity is presented for incomplete decision systems. Thirdly, a novel method for enhancing mutual granularity is proposed, which takes into account both the independence and correlation among candidate and selected attributes, with the aim of quantifying the uncertainty inherent in incomplete decision systems. Fourthly, an attribute selection algorithm based on enhanced mutual granularity is proposed. Finally, experimental results show that the proposed attribute selection method can effectively select the more relevant attributes with lower redundancy, thereby demonstrating strong classification capabilities when applied to incomplete decision systems.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Algorithm 1
Algorithm 2
Fig. 6

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

Data Availability

The data that support the findings of this study are available from the corresponding author upon reasonable request.

References

  1. Battiti R (1994) Using mutual information for selecting features in supervised neural net learning. IEEE Trans Neural Networks 5(4):537–550

    MATH  Google Scholar 

  2. Chen BW, Zhang XY, Yuan Z (2024) Two-dimensional improved attribute reductions based on distance granulation and condition entropy in incomplete interval-valued decision systems. Inf Sci 657:119910

    MATH  Google Scholar 

  3. Chen G, Chen J (2015) A novel wrapper method for feature selection and its applications. Neurocomputing 159:219–226

    MATH  Google Scholar 

  4. Chen HM, Li TR, Luo C, Horng SJ, Wang GY (2015) A decision-theoretic rough set approach for dynamic data mining. IEEE Trans Fuzzy Syst 23(6):1958–1970

    MATH  Google Scholar 

  5. Chen Z, Liu KY, Yang XB, Fujita H (2022) Random sampling accelerator for attribute reduction. Int J Approximate Reasoning 140:75–91

    MathSciNet  MATH  Google Scholar 

  6. Dai JH, Chen JL (2020) Feature selection via normative fuzzy information weight with application into tumor classification. Appl Soft Comput 92:106299

    MATH  Google Scholar 

  7. Dai JH, Tian HW (2013) Entropy measures and granularity measures for set-valued information systems. Inf Sci 240:72–82

    MathSciNet  MATH  Google Scholar 

  8. Dai JH, Xu Q (2012) Approximations and uncertainty measures in incomplete information systems. Inf Sci 198:62–80

    MathSciNet  MATH  Google Scholar 

  9. Dai JH, Wang WT, Xu Q (2012) An uncertainty measure for incomplete decision tables and its applications. IEEE Transactions on Cybernetics 43(4):1277–1289

    MATH  Google Scholar 

  10. Dai JH, Xu Q, Wang WT, Tian HW (2012) Conditional entropy for incomplete decision systems and its application in data mining. Int J Gen Syst 41(7):713–728

    MathSciNet  MATH  Google Scholar 

  11. Dai JH, Zhu ZL, Zou XT (2024) Fuzzy rough attribute reduction based on fuzzy implication granularity information. IEEE Trans Fuzzy Syst 32(6):3741–3752

    MATH  Google Scholar 

  12. Fleuret F (2004) Fast binary feature selection with conditional mutual information. J Mach Learn Res 5(9):1531–1555

    MathSciNet  MATH  Google Scholar 

  13. Frénay B, Doquire G, Verleysen M (2013) Theoretical and empirical study on the potential inadequacy of mutual information for feature selection in classification. Neurocomputing 112:64–78

    MATH  Google Scholar 

  14. Gu XY, Guo JC, Xiao LJ, Li CY (2022) Conditional mutual information-based feature selection algorithm for maximal relevance minimal redundancy. Appl Intell 52(2):1436–1447

    MATH  Google Scholar 

  15. Hancer E, Xue B, Zhang MJ (2018) Differential evolution for filter feature selection based on information theory and feature ranking. Knowl-Based Syst 140:103–119

    MATH  Google Scholar 

  16. Hoque N, Bhattacharyya DK, Kalita JK (2014) MIFS-ND: A mutual information-based feature selection method. Expert Syst Appl 41(14):6371–6385

    MATH  Google Scholar 

  17. Hu QH, Zhang L, Zhang D, Pan W, An S, Pedrycz W (2011) Measuring relevance between discrete and continuous features based on neighborhood mutual information. Expert Syst Appl 38(9):10737–10750

  18. Iyer R, Khargonkar N, Bilmes J, Asnani H (2021) Generalized submodular information measures: Theoretical properties, examples, optimization algorithms, and applications. IEEE Trans Inf Theory 68(2):752–781

    MathSciNet  MATH  Google Scholar 

  19. Kryszkiewicz M (1998) Rough set approach to incomplete information systems. Inf Sci 112(1–4):39–49

    MathSciNet  MATH  Google Scholar 

  20. Kwak N, Choi CH (2002) Input feature selection by mutual information based on parzen window. IEEE Trans Pattern Anal Mach Intell 24(12):1667–1671

    MATH  Google Scholar 

  21. Lewis DD (1992) Feature selection and feature extraction for text categorization. In: Speech and Natural Language: Proceedings of a Workshop Held at Harriman, New York, February 23-26, 1992, pp 212–217

  22. Li JL, Liu ZF (2024) Attribute-weighted outlier detection for mixed data based on parallel mutual information. Expert Syst Appl 236:121304

    MATH  Google Scholar 

  23. Liang JY, Shi ZZ (2004) The information entropy, rough entropy and knowledge granulation in rough set theory. Internat J Uncertain Fuzziness Knowledge-Based Systems 12(01):37–46

    MathSciNet  MATH  Google Scholar 

  24. Liang JY, Shi ZZ, Li DY, Wierman MJ (2006) Information entropy, rough entropy and knowledge granulation in incomplete information systems. Int J Gen Syst 35(6):641–654

    MathSciNet  MATH  Google Scholar 

  25. Liu HY, Zhou MC, Liu Q (2019) An embedded feature selection method for imbalanced data classification. IEEE/CAA Journal of Automatica Sinica 6(3):703–715

    MATH  Google Scholar 

  26. Liu KY, Li TR, Yang XB, Yang X, Liu D (2022) Neighborhood rough set based ensemble feature selection with cross-class sample granulation. Appl Soft Comput 131:109747

    MATH  Google Scholar 

  27. Liu KY, Li TR, Yang XB, Ju HR, Yang X, Liu D (2023) Feature selection in threes: Neighborhood relevancy, redundancy, and granularity interactivity. Appl Soft Comput 146:110679

    MATH  Google Scholar 

  28. Luo C, Li TR, Huang YY, Fujita H (2019) Updating three-way decisions in incomplete multi-scale information systems. Inf Sci 476:274–289

    MATH  Google Scholar 

  29. Luo JF, Fujita H, Yao YY, Qin KY (2020) On modeling similarity and three-way decision under incomplete information in rough set theory. Knowl-Based Syst 191:105251

    MATH  Google Scholar 

  30. Pawlak Z (1982) Rough sets. International Journal of Computer & Information Sciences 11(5):341–356

    MathSciNet  MATH  Google Scholar 

  31. Peng HC, Long FH, Ding C (2005) Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans Pattern Anal Mach Intell 27(8):1226–1238

    MATH  Google Scholar 

  32. Qian DM, Liu KY, Zhang SM, Yang XB (2024) Semi-supervised feature selection by minimum neighborhood redundancy and maximum neighborhood relevancy. Appl Intell 54:7750–7764

    MATH  Google Scholar 

  33. Qiang MZ, Shi ZZ (2009) A fast approach to attribute reduction in incomplete decision systems with tolerance relation-based rough sets. Inf Sci 179(16):2774–2793

    MathSciNet  MATH  Google Scholar 

  34. Slowinski R, Vanderpooten D (2000) A generalized definition of rough approximations based on similarity. IEEE Trans Knowl Data Eng 12(2):331–336

    MATH  Google Scholar 

  35. Sun L, Xu JC, Tian Y (2012) Feature selection using rough entropy-based uncertainty measures in incomplete decision systems. Knowl-Based Syst 36:206–216

  36. Sun L, Wang LY, Ding WP, Qian YH, Xu JC (2020) Neighborhood multi-granulation rough sets-based attribute reduction using lebesgue and entropy measures in incomplete neighborhood decision systems. Knowl-Based Syst 192:105373

  37. Swiniarski RW, Skowron A (2003) Rough set methods in feature selection and recognition. Pattern Recogn Lett 24(6):833–849

  38. Tawhid MA, Ibrahim AM (2020) Feature selection based on rough set approach, wrapper approach, and binary whale optimization algorithm. Int J Mach Learn Cybern 11(3):573–602

    MATH  Google Scholar 

  39. Thuy NN, Wongthanavasu S (2022) Hybrid filter-wrapper attribute selection with alpha-level fuzzy rough sets. Expert Syst Appl 193:116428

    MATH  Google Scholar 

  40. Vinh LT, Lee S, Park YT, d Auriol BJ (2012) A novel feature selection method based on normalized mutual information. Appl Intell 37(1):100–120

  41. Wang GQ, Li TR, Zhang PF, Huang QQ, Chen HM (2021) Double-local rough sets for efficient data mining. Inf Sci 571:475–498

    MathSciNet  MATH  Google Scholar 

  42. Wang J, Wei JM, Yang ZL, Wang SQ (2017) Feature selection by maximizing independent classification information. IEEE Trans Knowl Data Eng 29(4):828–841

    MATH  Google Scholar 

  43. Wang Q, Zhang XY (2024) Feature selection using three-stage heuristic measures based on mutual fuzzy granularities. Appl Intell 54(2):1445–1473

    MATH  Google Scholar 

  44. Wu SZ, Wang LT, Ge SY, Xiong Z, Liu J (2024) Feature selection algorithm using neighborhood equivalence tolerance relation for incomplete decision systems. Appl Soft Comput 157:111463

    MATH  Google Scholar 

  45. Yan T, Han CZ (2017) Entropy based attribute reduction approach for incomplete decision table. In: 2017 20th International Conference on Information Fusion (Fusion), IEEE, pp 1–8

  46. Yang XB, Zhang M, Dou HL, Yang JY (2011) Neighborhood systems-based rough sets in incomplete information system. Knowl-Based Syst 24(6):858–867

    MATH  Google Scholar 

  47. Zadeh LA (1979) Fuzzy sets and information granularity. Advances in Fuzzy Set Theory and Applications 11:3–18

    MathSciNet  MATH  Google Scholar 

  48. Zhang CC, Dai JH, Chen JL (2020) Knowledge granularity based incremental attribute reduction for incomplete decision systems. Int J Mach Learn Cybern 11(5):1141–1157

    MATH  Google Scholar 

  49. Zhang XY, Hou JL, Li JR (2022) Multigranulation rough set methods and applications based on neighborhood dominance relation in intuitionistic fuzzy datasets. Int J Fuzzy Syst 24(8):3602–3625

    MATH  Google Scholar 

  50. Zhang XY, Yuan Z, Miao DQ (2024) Outlier detection using three-way neighborhood characteristic regions and corresponding fusion measurement. IEEE Trans Knowl Data Eng 36(5):2082–2095

    MATH  Google Scholar 

  51. Zhou HF, Wang XQ, Zhu RR (2022) Feature selection based on mutual information with correlation coefficient. Appl Intell 52(5):5457–5474

    MATH  Google Scholar 

Download references

Acknowledgements

This work is supported by the National Natural Science Foundation of China (62376093, 61976089), the Major Program of the National Social Science Foundation of China (20 &ZD047), the Natural Science Foundation of Hunan Province (2021JJ30451, 2022JJ30397), the Hunan Provincial Science & Technology Project Foundation (2018RS3065, 2018TP1018), and the Postgraduate Scientific Research Innovation Project of Hunan Province (CX20240549).

Author information

Authors and Affiliations

Authors

Contributions

Conceptualization: Yongkang Zhang, Jianhua Dai; Methodology: Chucai Zhang; Writing-original draft preparation: Yongkang Zhang; Writing-review and editing: Jianhua Dai, Chucai Zhang.

Corresponding author

Correspondence to Jianhua Dai.

Ethics declarations

Competing Interests

The authors have no competing interests to declare that are relevant to the content of this article.

Ethical Approval

The data employed in this study are publicly accessible, with all usage conforming to standards of privacy protection and ethical guidelines.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zhang, C., Zhang, Y. & Dai, J. Attribute selection for incomplete decision systems by maximizing correlation and independence with mutual granularity. Appl Intell 55, 252 (2025). https://doi.org/10.1007/s10489-024-06170-x

Download citation

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s10489-024-06170-x

Keywords