Skip to main content
Log in

Noise cleaning for nonuniform ordinal labels based on inter-class distance

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

Label noise poses a significant challenge to supervised learning algorithms. Extensive research has been conducted on classification and regression tasks, but label noise filtering methods specifically designed for ordinal regression are lacking. In this paper, we propose a set of ordinal label noise filtering frameworks by theoretically exploring the generalization error bound in noisy environments. Besides, we present a robust label noise estimation method voted by inter-class distance. It takes into account the nonuniformity of ordinal labels and the reliability of the base model. This estimator is integrated into our framework in the proposed Inter-Class Distance-based Filtering (ICDF) algorithm. We empirically demonstrate the effectiveness of ICDF in identifying label noise and achieving improved generalization performance. Our experiments conducted on benchmark and real age estimation datasets show the superiority of ICDF over the existing filters in ordinal label noise cleaning.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Algorithm 1
Fig. 1
Fig. 2
Fig. 3
Fig. 4

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

Data availability and access

The benchmark ordinal regression datasets are available at http://www.gatsby.ucl.ac.uk/~chuwei/ordinalregression.html. The Adience age estimation dataset is provided at https://talhassner.github.io/home/projects/Adience/Adience-data.html.

References

  1. Singer G, Ratnovsky A, Naftali S (2021) Classification of severity of trachea stenosis from EEG signals using ordinal decision-tree based algorithms and ensemble-based ordinal and non-ordinal algorithms. Expert Syst Appl 173:114707

    Article  Google Scholar 

  2. Kamarajah S, Nepogodiev D, Bekele A et al (2021) Mortality from esophagectomy for esophageal cancer across low, middle, and high-income countries: an international cohort study. Eur J Surg Oncol 47(6):1481–1488

    Article  Google Scholar 

  3. Gutiérrez PA, Pérez-Ortiz M, Sánchez-Monedero J et al (2016) Ordinal regression methods: survey and experimental study. IEEE Trans Knowl Data Eng 28(1):127–146

    Article  Google Scholar 

  4. Cao W, Mirjalili V, Raschka S (2020) Rank consistent ordinal regression for neural networks with application to age estimation. Pattern Recogn Lett 140:325–331

    Article  Google Scholar 

  5. Tang MZ, Pérez-Fernández R, De Baets B (2022) Ordinal classification with a spectrum of information sources. Expert Syst Appl 208:118163

    Article  Google Scholar 

  6. Liu XF, Fan FF, Kong LS et al (2020) Unimodal regularized neuron stick-breaking for ordinal classification. Neurocomputing 388:34–44

    Article  Google Scholar 

  7. Shi J, Cao Z, Wu J (2022) Meta joint optimization: a holistic framework for noisy-labeled visual recognition. Appl Intell 52(1):875–888

    Article  Google Scholar 

  8. He GL, Li B, Wang H et al (2022) Cost-effective active semi-supervised learning on multivariate time series data with crowds. IEEE Transactions on Systems, Man, and Cybernetics: Systems 52(3):1437–1450

    Article  Google Scholar 

  9. Shu J, Yuan X, Meng D et al (2023) Cmw-net: Learning a class-aware sample weighting mapping for robust deep learning. IEEE Trans Pattern Anal Mach Intell 45(10):11521–11539

    Article  Google Scholar 

  10. Ma B, Li C, Jiang L (2022) A novel ground truth inference algorithm based on instance similarity for crowdsourcing learning. Appl Intell 52(15):17784–17796

    Article  Google Scholar 

  11. Sáez JA, Corchado E (2022) ANCES: A novel method to repair attribute noise in classification problems. Pattern Recogn 121:108198

    Article  Google Scholar 

  12. Wang K, Yang M, Yang W et al (2022) Dual-scale correlation analysis for robust multi-label classification. Appl Intell 52(14):16382–16397

    Article  Google Scholar 

  13. Rojarath A, Songpan W (2021) Cost-sensitive probability for weighted voting in an ensemble model for multi-class classification problems. Appl Intell 51:4908–4932

    Article  Google Scholar 

  14. Song H, Dai R, Raskutti G et al (2020) Convex and non-convex approaches for statistical inference with class-conditional noisy labels. J Mach Learn Res 21(168):1–58

    MathSciNet  Google Scholar 

  15. Song H, Kim M, Park D et al (2023) Learning from noisy labels with deep neural networks: A survey. IEEE Trans Neural Netw Learn Syst 34(11):8135–8153

    Article  Google Scholar 

  16. Leng Q, He Z, Liu Y et al (2021) A soft-margin convex polyhedron classifier for nonlinear task with noise tolerance. Appl Intell 51:453–466

    Article  Google Scholar 

  17. Van Engelen JE, Hoos HH (2020) A survey on semi-supervised learning. Mach Learn 109(2):373–440

    Article  MathSciNet  Google Scholar 

  18. Kim T, Ko J, Cho S et al (2021) FINE samples for learning with noisy labels. In: Advances in neural information processing systems, pp 24137–24149

  19. Liao T, Lei Z, Zhu T et al (2023) Deep metric learning for K nearest neighbor classification. IEEE Trans Knowl Data Eng 35(1):264–275

    Google Scholar 

  20. Eronen J, Ptaszynski M, Masui F et al (2021) Improving classifier training efficiency for automatic cyberbullying detection with feature density. Inform Process Manag 58(5):102616

    Article  Google Scholar 

  21. Kordos M, Blachnik M, Scherer R (2022) Fuzzy clustering decomposition of genetic algorithm-based instance selection for regression problems. Inf Sci 587:23–40

    Article  Google Scholar 

  22. Dong Y, Jiang LX, Li CQ (2022) Improving data and model quality in crowdsourcing using co-training-based noise correction. Inf Sci 583:174–188

    Article  Google Scholar 

  23. Li C, Mao Z (2023) A label noise filtering method for regression based on adaptive threshold and noise score. Expert Syst Appl 228:120422

    Article  Google Scholar 

  24. Chen Q, Jiang G, Cao F et al (2024) A general elevating framework for label noise filters. Pattern Recogn 147:110072

    Article  Google Scholar 

  25. Jiang GX, Wang WJ, Qian YH et al (2021) A unified sample selection framework for output noise filtering: An error-bound perspective. J Mach Learn Res 22(18):1–66

    MathSciNet  Google Scholar 

  26. Jiang GX, Wang WJ (2022) A numerical label noise filtering algorithm for regression task. J Comput Res Dev 59(8):1639–1652

    Google Scholar 

  27. Wang C, Shi J, Zhou YY et al (2022) Label noise modeling and correction via loss curve fitting for SAR ATR. IEEE Trans Geosci Remote Sens 60:1–10

  28. Yuan Y, Wan S, Zhang C et al (2023) CASSOR: Class-aware sample selection for ordinal regression with noisy labels. 20th Pacific rim international conference on artificial intelligence, Jakarta Indonesia. Accessed 15–19 Nov 2023

  29. Zhu F, Chen X, Gao X et al (2023) Constraint-weighted support vector ordinal regression to resist constraint noises. Inf Sci 649:119644

    Article  Google Scholar 

  30. Punyani P, Gupta R, Kumar A (2020) Neural networks for facial age estimation: a survey on recent advances. Artif Intell Rev 53:3299–3347

Download references

Funding

This work was supported by the National Natural Science Foundation of China (62276161, U21A20513, 62076154, 61906113), and the Key R &D Program of Shanxi Province (202202020101003, 202302010101007)

Author information

Authors and Affiliations

Authors

Contributions

All authors contributed to the algorithm conception and design. Theoretical analysis was completed by Gaoxia Jiang and Wenjian Wang. Data collection and analysis were performed by Fei Wang. The first draft of the manuscript was written by Gaoxia Jiang. All authors commented on previous versions of the manuscript and approved the final version.

Corresponding author

Correspondence to Gaoxia Jiang.

Ethics declarations

Ethical and informed consent for data used

The benchmark ordinal regression datasets are open source. The Adience age estimation dataset will be made available upon reasonable request.

Competing Interests

The authors have no relevant financial or nonfinancial interests to disclose.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Jiang, G., Wang, F. & Wang, W. Noise cleaning for nonuniform ordinal labels based on inter-class distance. Appl Intell 54, 6997–7011 (2024). https://doi.org/10.1007/s10489-024-05551-6

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10489-024-05551-6

Keywords