Abstract
Label noise poses a significant challenge to supervised learning algorithms. Extensive research has been conducted on classification and regression tasks, but label noise filtering methods specifically designed for ordinal regression are lacking. In this paper, we propose a set of ordinal label noise filtering frameworks by theoretically exploring the generalization error bound in noisy environments. Besides, we present a robust label noise estimation method voted by inter-class distance. It takes into account the nonuniformity of ordinal labels and the reliability of the base model. This estimator is integrated into our framework in the proposed Inter-Class Distance-based Filtering (ICDF) algorithm. We empirically demonstrate the effectiveness of ICDF in identifying label noise and achieving improved generalization performance. Our experiments conducted on benchmark and real age estimation datasets show the superiority of ICDF over the existing filters in ordinal label noise cleaning.





Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data availability and access
The benchmark ordinal regression datasets are available at http://www.gatsby.ucl.ac.uk/~chuwei/ordinalregression.html. The Adience age estimation dataset is provided at https://talhassner.github.io/home/projects/Adience/Adience-data.html.
References
Singer G, Ratnovsky A, Naftali S (2021) Classification of severity of trachea stenosis from EEG signals using ordinal decision-tree based algorithms and ensemble-based ordinal and non-ordinal algorithms. Expert Syst Appl 173:114707
Kamarajah S, Nepogodiev D, Bekele A et al (2021) Mortality from esophagectomy for esophageal cancer across low, middle, and high-income countries: an international cohort study. Eur J Surg Oncol 47(6):1481–1488
Gutiérrez PA, Pérez-Ortiz M, Sánchez-Monedero J et al (2016) Ordinal regression methods: survey and experimental study. IEEE Trans Knowl Data Eng 28(1):127–146
Cao W, Mirjalili V, Raschka S (2020) Rank consistent ordinal regression for neural networks with application to age estimation. Pattern Recogn Lett 140:325–331
Tang MZ, Pérez-Fernández R, De Baets B (2022) Ordinal classification with a spectrum of information sources. Expert Syst Appl 208:118163
Liu XF, Fan FF, Kong LS et al (2020) Unimodal regularized neuron stick-breaking for ordinal classification. Neurocomputing 388:34–44
Shi J, Cao Z, Wu J (2022) Meta joint optimization: a holistic framework for noisy-labeled visual recognition. Appl Intell 52(1):875–888
He GL, Li B, Wang H et al (2022) Cost-effective active semi-supervised learning on multivariate time series data with crowds. IEEE Transactions on Systems, Man, and Cybernetics: Systems 52(3):1437–1450
Shu J, Yuan X, Meng D et al (2023) Cmw-net: Learning a class-aware sample weighting mapping for robust deep learning. IEEE Trans Pattern Anal Mach Intell 45(10):11521–11539
Ma B, Li C, Jiang L (2022) A novel ground truth inference algorithm based on instance similarity for crowdsourcing learning. Appl Intell 52(15):17784–17796
Sáez JA, Corchado E (2022) ANCES: A novel method to repair attribute noise in classification problems. Pattern Recogn 121:108198
Wang K, Yang M, Yang W et al (2022) Dual-scale correlation analysis for robust multi-label classification. Appl Intell 52(14):16382–16397
Rojarath A, Songpan W (2021) Cost-sensitive probability for weighted voting in an ensemble model for multi-class classification problems. Appl Intell 51:4908–4932
Song H, Dai R, Raskutti G et al (2020) Convex and non-convex approaches for statistical inference with class-conditional noisy labels. J Mach Learn Res 21(168):1–58
Song H, Kim M, Park D et al (2023) Learning from noisy labels with deep neural networks: A survey. IEEE Trans Neural Netw Learn Syst 34(11):8135–8153
Leng Q, He Z, Liu Y et al (2021) A soft-margin convex polyhedron classifier for nonlinear task with noise tolerance. Appl Intell 51:453–466
Van Engelen JE, Hoos HH (2020) A survey on semi-supervised learning. Mach Learn 109(2):373–440
Kim T, Ko J, Cho S et al (2021) FINE samples for learning with noisy labels. In: Advances in neural information processing systems, pp 24137–24149
Liao T, Lei Z, Zhu T et al (2023) Deep metric learning for K nearest neighbor classification. IEEE Trans Knowl Data Eng 35(1):264–275
Eronen J, Ptaszynski M, Masui F et al (2021) Improving classifier training efficiency for automatic cyberbullying detection with feature density. Inform Process Manag 58(5):102616
Kordos M, Blachnik M, Scherer R (2022) Fuzzy clustering decomposition of genetic algorithm-based instance selection for regression problems. Inf Sci 587:23–40
Dong Y, Jiang LX, Li CQ (2022) Improving data and model quality in crowdsourcing using co-training-based noise correction. Inf Sci 583:174–188
Li C, Mao Z (2023) A label noise filtering method for regression based on adaptive threshold and noise score. Expert Syst Appl 228:120422
Chen Q, Jiang G, Cao F et al (2024) A general elevating framework for label noise filters. Pattern Recogn 147:110072
Jiang GX, Wang WJ, Qian YH et al (2021) A unified sample selection framework for output noise filtering: An error-bound perspective. J Mach Learn Res 22(18):1–66
Jiang GX, Wang WJ (2022) A numerical label noise filtering algorithm for regression task. J Comput Res Dev 59(8):1639–1652
Wang C, Shi J, Zhou YY et al (2022) Label noise modeling and correction via loss curve fitting for SAR ATR. IEEE Trans Geosci Remote Sens 60:1–10
Yuan Y, Wan S, Zhang C et al (2023) CASSOR: Class-aware sample selection for ordinal regression with noisy labels. 20th Pacific rim international conference on artificial intelligence, Jakarta Indonesia. Accessed 15–19 Nov 2023
Zhu F, Chen X, Gao X et al (2023) Constraint-weighted support vector ordinal regression to resist constraint noises. Inf Sci 649:119644
Punyani P, Gupta R, Kumar A (2020) Neural networks for facial age estimation: a survey on recent advances. Artif Intell Rev 53:3299–3347
Funding
This work was supported by the National Natural Science Foundation of China (62276161, U21A20513, 62076154, 61906113), and the Key R &D Program of Shanxi Province (202202020101003, 202302010101007)
Author information
Authors and Affiliations
Contributions
All authors contributed to the algorithm conception and design. Theoretical analysis was completed by Gaoxia Jiang and Wenjian Wang. Data collection and analysis were performed by Fei Wang. The first draft of the manuscript was written by Gaoxia Jiang. All authors commented on previous versions of the manuscript and approved the final version.
Corresponding author
Ethics declarations
Ethical and informed consent for data used
The benchmark ordinal regression datasets are open source. The Adience age estimation dataset will be made available upon reasonable request.
Competing Interests
The authors have no relevant financial or nonfinancial interests to disclose.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Jiang, G., Wang, F. & Wang, W. Noise cleaning for nonuniform ordinal labels based on inter-class distance. Appl Intell 54, 6997–7011 (2024). https://doi.org/10.1007/s10489-024-05551-6
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-024-05551-6