Abstract
As an effective method for mining latent information between labels, label correlation is widely adopted by many scholars to model multi-label learning algorithms. Most existing multi-label algorithms usually ignore that the correlation between labels may be asymmetric while asymmetry correlation commonly exists in the real-world scenario. To tackle this problem, a multi-label learning algorithm with asymmetry label correlation (ACML, Asymmetry Label Correlation for Multi-Label Learning) is proposed in this paper. First, measure the adjacency between labels to construct the label adjacency matrix. Then, cosine similarity is utilized to construct the label correlation matrix. Finally, we constrain the label correlation matrix with the label adjacency matrix. Thus, asymmetry label correlation is modeled for multi-label learning. Experiments on multiple multi-label benchmark datasets show that the ACML algorithm has certain advantages over other comparison algorithms. The results of statistical hypothesis testing further illustrate the effectiveness of the proposed algorithm.
Similar content being viewed by others
References
Zhang ML, Zhou ZH (2013) A review on multi-label learning algorithms[J]. IEEE Trans Knowl Data Eng 26(8):1819–1837
Tsoumakas G, Katakis I (2007) Multi-label classification: an overview[J]. Int J Data Warehousing Min 3(3):1–13
Yu Y, Pedrycz W, Miao D (2014) Multi-label classification by exploiting label correlations[J]. Expert Syst Appl 41(6):2989–3004
Xue L, Jiang D, Wang R (2020) Learning semantic dependencies with channel correlation for multi-label classification[J]. Vis Comput 36(7):1325–1335
Han H, Huang M, Zhang Y, Yang X, Feng W (2019) Multi-label learning with label specific features using correlation information[J]. IEEE Access 7:11474–11484
Zhu Y, Kwok JT, Zhou ZH (2017) Multi-label learning with global and local label correlation[J]. IEEE Trans Knowl Data Eng 30(6):1081–1094
Cheng YS, Qian K, Wang YB, Zhao DW (2019) Multi-label lazy learning approach based on firefly method[J]. J Comput Appl 039(005):1305–1311
Huang SJ, Yu Y, Zhou ZH (2012) Multi-label hypothesis reuse[A]. Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining[C]. 525–533
Zhang M L, Zhang K (2010) Multi-label learning by exploiting label dependency[A]. Proceedings of the 16th ACM SIGKDD Conference on Knowledge Discovery and Data Mining[C]. Washington D. C., 999–1007
Székely GJ, Rizzo ML, Bakirov NK (2007) Measuring and testing dependence by correlation of distances. Ann Stat 35(6):2769–2794
Zhou Y, Zhang Y, Zhu L (2020) A projective approach to conditional independence test for dependent processes[J]. J Bus Econ Stat. https://doi.org/10.1080/07350015.2020.1826952
Srinivasan G, Hyman JD, Osthus DA et al (2018) Quantifying topological uncertainty in fractured systems using graph theory and machine learning[J]. Sci Rep 8(1):1–11
Li J, Li X, He D (2019) A directed acyclic graph network combined with CNN and LSTM for remaining useful life prediction[J]. IEEE Access 7:75464–75475
Mihaljević B, Bielza C, Larrañaga P (2018) Learning Bayesian network classifiers with completed partially directed acyclic graphs[C]//international conference on probabilistic graphical models. PMLR:272–283
Liu F, Chan L (2016) Causal inference on discrete data via estimating distance correlations[J]. Neural Comput 28(5):801–814
Beck A, Teboulle M (2009) Fast gradient-based algorithms for constrained total variation image denoising and deblurring problems[J]. IEEE Trans Image Process 18(11):2419–2434
Ganesh A, Lin Z, Wright J, et al. (2009) Fast algorithms for recovering a corrupted low-rank matrix[C]//2009 3rd IEEE international workshop on computational advances in multi-sensor adaptive processing (CAMSAP). IEEE 213–216
Huang J, Qin F, Zheng X, Cheng Z, Yuan Z, Zhang W, Huang Q (2019) Improving multi-label classification with missing labels by learning label-specific features[J]. Inf Sci 492:124–146
Gibaja EL, Ventura S (2015) A tutorial on multi-label learning[J]. ACM Comput Surv (CSUR) 47(3):1–38
Sorower MS (2010) A literature survey on algorithms for multi-label learning[J]. Oregon State University. https://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.364.5612&rep=rep1&type=pdf
Tsoumakas G, Katakis I, Vlahavas I (2009) Mining multi-label data[M]. In: Data Mining and Knowledge Discovery Handbook. Springer, Boston. https://doi.org/10.1007/978-0-387-09823-4_34
Huang J, Li G, Huang Q, Wu X (2015) Learning Label Specific Features for Multi-Label Classification[C]// 2015 IEEE International Conference on Data Mining. IEEE 181–190
Kumar V, Pujari AK, Padmanabhan V, Kagita VR (2019) Group preserving label embedding for multi-label classification[J]. Pattern Recogn 90:23–34
Acknowledgments
This work was supported in part by the National Natural Science Foundation of China under Grant 61702012 and Key Laboratory of Data Science and Intelligence Application, Fujian Province University (NO. D202005) and Key Laboratory of Intelligent Computing & Signal Processing, Ministry of Education (Anhui University) (No.2020A003).
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Bao, J., Wang, Y. & Cheng, Y. Asymmetry label correlation for multi-label learning. Appl Intell 52, 6093–6105 (2022). https://doi.org/10.1007/s10489-021-02725-4
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-021-02725-4