Skip to main content
Log in

Measures of uncertainty for partially labeled categorical data based on an indiscernibility relation: an application in semi-supervised attribute reduction

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

In many practical applications of machine learning, only part of data is labeled because the cost of assessing class label is relatively high. This paper concentrates on measures of uncertainty for a partial label categorical decision information system (p-CDIS), and considers an application to semi-supervised attribute reduction. Firstly, two decision information systems (DISs) can be induced by a p-CDIS (UCd): one is for a decision information system for labeled categorical data \((U^l,C,d)\) and the other one is a decision information system for unlabeled categorical data \((U^u,C,d)\), and the missing rate of labels in (UCd) is introduced. In view of partial label data, the existential research did not take into account the missing rate of labels and only considered one importance of each attribute subset. Then, four importance of an attribute subset \(P\subseteq C\) in (UCd) are defined based on an indiscernibility relation. They are the weighted sum of the importance of P in \((U^l,C,d)\) and \((U^u,C,d)\) determined by the missing rate of labels. These four importance can be regarded as four uncertainty measurements (UMs) for (UPd). Next, numerical experiments and statistical tests are carried out on 15 datasets of UCI to demonstrate four UMs’ advantages and disadvantages. Finally, as an application for UM in p-CDIS, two better UMs are used as semi-supervised attribute reduction and two corresponding algorithms are designed that can automatically adapt to different missing rates of labels. The experimental results show the feasibility and superiority of the designed algorithms.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Algorithm 1
Algorithm 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8

Similar content being viewed by others

Data availability and access

The data used or analyzed during the current study are available from the corresponding author after the paper is accepted for publication.

References

  1. Andrzej C (2003) Automatic identication of sound source position employing neural networks and rough sets. Pattern Recognit Lett 24(6):921–933

    Google Scholar 

  2. Benabdeslem K, Hindawi M (2014) Efficient semi-supervised feature selection: constraint, relevance and redundancy. IEEE Trans Knowl Data Eng 26:1131–1143

    Google Scholar 

  3. Bao WX, Hang JY, Zhang ML (2021) Partial label dimensionality reduction via confidence-based dependence maximization. In: Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, pp. 46–54

  4. Bao WX, Hang JY, Zhang ML (2022) Submodular feature selection for partial label learning. In: Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp. 26–34

  5. Campagner A, Ciucci D (2022) Rough-set based genetic algorithms for weakly supervised feature selection. International Conference on Information Processing and Management of Uncertainty in Knowledge-Based Systems. Springer, Cham, pp 761–773

    Google Scholar 

  6. Campagner A, Ciucci D, Huellermeier E (2021) Rough set-based feature selection for weakly labeled data. Int J Approx Reason 136:150–167

    MathSciNet  MATH  Google Scholar 

  7. Blum A, Mitchell T (1998) Combining labeled and unlabeled data with cortraining//Proceedings of the 11th annual Conference on Computational Learning Theory. New York: ACM, 92–100

  8. Beaubouef T, Petry FE, Arora G (1998) Information-theoretic measures of uncertainty for rough sets and rough relational databases. Inf Sci 109:185–195

    Google Scholar 

  9. Chen YQ, Gao W, Zhu TS (2002) Learning prosodic patterns for mandarin speech synthesis. J Intell Inf Syst 19(1):95–109

    Google Scholar 

  10. Chen Z, Liu KY, Yang XB, Fujita H (2022) Random sampling accelerator for attribute reduction. Int J Approx Reason 140:75–91

    MathSciNet  MATH  Google Scholar 

  11. Chen Y, Liu KY, Song JJ, Fujita H, Yang XB, Qian YH (2020) Attribute group for attribute reduction. Inf Sci 535:64–80

    MATH  Google Scholar 

  12. Girish C, Ferat S (2014) A survey on feature selection methods. Comput Electr Eng 40:16–28

    Google Scholar 

  13. Dai JH, Hu QH, Zhang JH, Hu H, Zheng NG (2017) Attribute selection for partially labeled categorical data by rough set approach. IEEE Trans Cybernet 47:2460–2471

    Google Scholar 

  14. Dai JH, Wang WS, Zhang CC, Qu SJ (2023) Semi-supervised attribute reduction via attribute indiscernibility. Int J Mach Learn Cybernet 14(4):1445–1464

    Google Scholar 

  15. Feofanov V, Devijver E, Amini MR (2022) Wrapper feature selection with partially labeled data. Appl Intell 52(11):12316–12329

    Google Scholar 

  16. Fan TF, Liu DR, Tzeng GH (2007) Rough set-based logics for multicriteria decision analysis. European J Operat Res 182:340–355

    MathSciNet  MATH  Google Scholar 

  17. Forestier G, Wemmert C (2016) Semi-supervised learning using multiple clustering with limited labeled data. Inf Sci 361–362:48–65

    Google Scholar 

  18. He XF, Deng C, Partha N (2005) Laplacian score for feature selection//Proceedings of the 18th International Conference on Neural Information Processing Systems (NIPS’05). Cambridge, USA: MIT Press, 507–514

  19. Handl J, Knowles J (2006) Semi-supervised feature selection via multi-objective optimization//The 2006 International Joint Conference on Neural Networks

  20. Hirano S, Tsumoto S (2005) Rough representation of a region of interest in medical images. Int J Approx Reason 40(1–2):23–34

    Google Scholar 

  21. Han YH, Yang Y, Yan Y, Ma ZG, Zhou XF (2015) Semisupervised feature selection via spline regression for video semantic recognition. IEEE Trans Neural Netw Learn Syst 26:252–264

    MathSciNet  Google Scholar 

  22. Weston J, Andre E, Bernhard S (2003) Use of the zero-norm with linear models and kernel methods. J Mach Learn Res 3:1439–1461

    MathSciNet  MATH  Google Scholar 

  23. Jiang ZH, Liu KY, Yang XB, Yu HL, Fujita H, Qian YH (2020) Accelerator for supervised neighborhood based attribute reduction. Int J Approx Reason 119:122–150

    MathSciNet  MATH  Google Scholar 

  24. Jiang F, Sui YF, Cao CG (2010) An information entropy-based approach to outlier detection in rough sets. Exp Syst Appl 37:6338–6344

    Google Scholar 

  25. Jain P, Tiwari AK, Som T (2020) A fitting model based intuitionistic fuzzy rough feature selection. Eng Appl Art Intell 89:103421

    Google Scholar 

  26. Kohavi R, John GH (1997) Wrappers for feature subset selection. Art Intell 97(1–2):273–324

    MATH  Google Scholar 

  27. Liu GL (2022) Attribute reduction algorithms determined by invariants for decision tables. Cognit Comput 14:1818–1825

    Google Scholar 

  28. Liang JY, Shi ZZ, Li DY (2006) Information entropy, rough entropy and knowledge granulation in incomplete information systems. Int J Gen Syst 35:641–654

    MathSciNet  MATH  Google Scholar 

  29. Liu KY, Yang XB, Yu HL, Mi JS (2019) Rough set based semi-supervised feature selection via ensemble selector. Knowl-Based Syst 165:282–296

    Google Scholar 

  30. Kryszkiewicz M (1999) Rules in incomplete information systems. Inf Sci 113:271–292

    MathSciNet  MATH  Google Scholar 

  31. Miao D, Gao C, Zhang N (2011) Diverse reduct subspaces based co-training for partially labeled data. Int J Approx Reason 52:1103–1117

    MathSciNet  Google Scholar 

  32. Nakatani Y, Zhu K, Uehara K (2007) Semi-supervised learning using feature selection based on maximum density subgraphs. Syst Comput Japan 38:32–43

    Google Scholar 

  33. Pawlak Z (1982) Rough sets. Int J Comput Inf Sci 11:341–356

    MATH  Google Scholar 

  34. Qian YH, Liang JY (2008) Combination entropy and combination granulation in rough set theory. Int J Uncert Fuzz Knowl-Based Syst 16:179–193

    MathSciNet  MATH  Google Scholar 

  35. Ren JY, Qiu ZY, Fan W (2008) Forward semi-supervised feature selectio. in: Proceedings of the 12th Pacific-Asia conference on Advances in Knowledge Discovery and Data Mining (PAKDD’08). Springer-Verlag, Berlin, pp. 970–976

  36. Shannon C (1948) A mathematical theory of communication. Bell Syst Tech J 27:379–423

    MathSciNet  MATH  Google Scholar 

  37. Sang BB, Chen HM, Yang L, Li TR, Xu WH (2021) Incremental feature selection using a conditional entropy based on fuzzy dominance neighborhood rough sets. IEEE Trans Fuzz Syst 30(6):1683–1697

    Google Scholar 

  38. UCI Machine Learning Repository. http://archive.ics.uci.edu/ml/datasets.html

  39. Wierman MJ (1999) Measuring uncertainty in rough set theory. Int J Gen Syst 28:283–297

    MathSciNet  MATH  Google Scholar 

  40. Wang XZ, Tsang ECC, Zhao SY, Chen DG, Yeung DS (2007) Learning fuzzy rules from fuzzy samples based on rough set technique. Inf Sci 177:4493–4514

    MathSciNet  MATH  Google Scholar 

  41. Wang CZ, Wang Y, Shao MW, Qian YH, Chen DG (2020) Fuzzy rough attribute reduction for categorical data. IEEE Trans Fuzz Syst 28:818–830

    Google Scholar 

  42. Wan L, Xia SJ, Zhu Y, Lyu ZH (2021) An improved semi-supervised feature selection algorithm based on information entropy. Stat Decis Making 17:66–70

    Google Scholar 

  43. Wang YB, Chen XJ, Dong K (2019) Attribute reduction via local conditional entropy. Int J Mach Learn Cybernet 10:3619–3634

    Google Scholar 

  44. Yuan Z, Chen HM, Zhang PF, Wan JH, Li TR (2021) A novel unsupervised approach to heterogeneous feature selection based on fuzzy mutual information. IEEE Trans Fuzz Syst 30(9):3395–3409

    Google Scholar 

  45. Yang X, Chen Y, Fujita H, Liu D, Li TR (2022) Mixed data-driven sequential three-way decision via subjective-objective dynamic fusion. Knowl-Based Syst 237:107728

    Google Scholar 

  46. Zhao MY, Jiao LC, Ma WP (2016) Classification and saliency detection by semi-supervised low-rank representation. Pattern Recognit 51:281–294

    Google Scholar 

  47. Zhang W, Miao DQ, Gao C, Li F (2016) Semi-supervised attribute reduction based on rough-subspace ensemble learning. J Chinese Comput Syst 37:2727–2732

    Google Scholar 

  48. Zhang R, Nie F, Li X (2019) Feature selection with multi-view data: A survey. Inf Fusion 50:158–167

    Google Scholar 

  49. Hu X, Zhou P, Li P, Wang J, Wu X (2018) A survey on online feature selection with streaming features. Front Comput Sci 12:479–493

    Google Scholar 

Download references

Acknowledgements

The authors would like to thank the editors and the anonymous reviewers for their valuable comments and suggestions, which have helped immensely in improving the quality of the paper. This work is supported by Natural Science Foundation of Guangxi Province (2021GXNSFAA220114, 2020GXNSFAA159155) and Guangxi First-class Discipline Statistics Construction Project Fund.

Author information

Authors and Affiliations

Authors

Contributions

Jiali He: Methodology, Writing-Original draft; Gangqiang Zhang: Software, Editing, Investigation; Dan Huang: Data curation; Pei Wang: Validation; Guangji Yu: Software, Investigation.

Corresponding authors

Correspondence to Gangqiang Zhang or Guangji Yu.

Ethics declarations

Competing Interests

The authors declare that they have no conflict of interest.

Ethical and informed consent for data used

The data used or analyzed during the current study are available from the corresponding author after the paper is accepted for publication.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

He, J., Zhang, G., Huang, D. et al. Measures of uncertainty for partially labeled categorical data based on an indiscernibility relation: an application in semi-supervised attribute reduction. Appl Intell 53, 29486–29513 (2023). https://doi.org/10.1007/s10489-023-05078-2

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10489-023-05078-2

Keywords

Navigation