Abstract
Multi-label learning has attracted many attentions. However, the continuous data generated in the fields of sensors, network access, etc., that is data streams, the scenario brings challenges such as real-time, limited memory, once pass. Several learning algorithms have been proposed for offline multi-label classification, but few researches develop it for dynamic multi-label incremental learning models based on cascading schemes. Deep forest can perform representation learning layer by layer, and does not rely on backpropagation, using this cascading scheme, this paper proposes a multi-label data stream deep forest (VDSDF) learning algorithm based on cascaded Very Fast Decision Tree (VFDT) forest, which can receive examples successively, perform incremental learning, and adapt to concept drift. Experimental results show that the proposed VDSDF algorithm, as an incremental classification algorithm, is more competitive than batch classification algorithms on multiple indicators. Moreover, in dynamic flow scenarios, the adaptability of VDSDF to concept drift is better than that of the contrast algorithm.
Similar content being viewed by others
References
Liu W, Wang H, Shen X, Tsang I (2021) The emerging trends of multi-label learning. IEEE Transactions on Pattern Analysis and Machine Intelligence
Balakrishna S, Thirumaran M, Solanki VK, Núñez Valdéz ER, et al. (2020) Incremental hierarchical clustering driven automatic annotations for unifying iot streaming data. International Journal Of Interactive Multimedia And Artificial Intelligence
Moyano JM, Gibaja EL, Cios KJ, Ventura S (2018) Review of ensembles of multi-label classifiers: models, experimental study and prospects. Information Fusion 44:33–45
Yan M (2020) Accurate detecting concept drift in evolving data streams. ICT Express
Cerri R, Júnior JDC, Paiva ERF, da Gama JMP (2020) Multi-label stream classification with self-organizing maps. arXiv preprint arXiv:2004.09397
Wang L, Lin X, Lin N (2021) Research on pseudo-label technology for multi-label news classification. In: International conference on document analysis and recognition, Springer, pp 683–698
Cerri R, Basgalupp MP, Barros RC, de Carvalho ACPLF (2019) Inducing hierarchical multi-label classification rules with genetic algorithms. Appl Soft Comput
Pliakos K, Geurts P, Vens C (2018) Global multi-output decision trees for interaction prediction. Mach Learn 107(8):1257–1281
Wang L, Liu Y, Qin C, Sun G, Fu Y (2020) Dual relation semi-supervised multi-label learning. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol 34, pp 6227–6234
Gong X, Yuan D, Bao W (2020) Online metric learning for multi-label classification. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol 34, pp 4012–4019
Mehta S et al (2017) Concept drift in streaming data classification: algorithms, platforms and issues. Procedia computer science 122:804–811
Yang L, Wu X-Z, Jiang Y, Zhou Z-H (2019) Multi-label learning with deep forest. arXiv:1911.06557
Yapp EKY, Li X, Lu WF, Tan PS (2020) Comparison of base classifiers for multi-label learning. Neurocomputing 394:51–60
Zhang C, Li Z (2021) Multi-label learning with label-specific features via weighting and label entropy guided clustering ensemble. Neurocomputing 419:59–69
Rodrigues J, Cristo M, Colonna JG (2020) Deep hashing for multi-label image retrieval: a survey. Artif Intell Rev 53(7):5261–5307
Gomes HM, Bifet A, Read J, Barddal JP, Enembreck F, Pfharinger B, Holmes G, Abdessalem T (2017) Adaptive random forests for evolving data stream classification. Mach Learn 106(9):1469–1495
Wang H, Abraham Z (2015) Concept drift detection for streaming data. In: 2015 international joint conference on neural networks (IJCNN), IEEE, pp 1–9
Luong AV, Nguyen TT, Liew A W-C (2020) Streaming active deep forest for evolving data stream classification. arXiv:2002.11816
Wang W, Dai Q, Li F, Xiong Y, Wei D-Q (2021) Mlcdforest: multi-label classification with deep forest in disease prediction for long non-coding rnas. Brief Bioinform 22(3):bbaa104
Zhou Z-H, Feng J (2017) Deep forest: Towards an alternative to deep neural networks. arXiv:1702.08835
Gibaja E, Ventura S (2014) Multi-label learning: a review of the state of the art and ongoing research. Wiley Interdisciplinary Reviews: Data Knowl Eng 4(6):411–444
Read J, Pfahringer B, Holmes G, Frank E (2011) Classifier chains for multi-label classification. Mach Learn 85(3):333
Zhang M-L, Zhou Z-H (2013) A review on multi-label learning algorithms. IEEE Trans Knowl Data Eng 26(8):1819–1837
Zhang M-L, Zhou Z-H (2007) Ml-knn: A lazy learning approach to multi-label learning. Pattern Recog 40(7):2038–2048
Yixiao H, Ming P, Yuan J (2020) Mondrian deep forest. J Comput Res Dev 57(8):1594
Sheng-Jun H, Wei G, Zhi-Hua Z (2020) Fast multi-instance multi-label learning. IEEE Transactions on Pattern Analysis and Machine Intelligence
Read J, Bifet A, Holmes G, Pfahringer B (2012) Scalable and efficient multi-label classification for evolving data streams. Mach Learn 88(1-2):243–272
Kocev D, Vens C, Struyf J, Džeroski S (2007) Ensembles of multi-objective decision trees. In: European conference on machine learning, Springer, pp 624–631
Garcıa-Mart?n E, Lavesson N, Grahn H, Casalicchio E, Boeva V (2018) Hoeffding trees with nmin adaptation. In: IEEE
Venkatesan R, Er MJ, Wu S, Pratama M (2016) A novel online real-time classifier for multi-label data streams. In: 2016 International Joint Conference on Neural Networks (IJCNN), IEEE, pp 1833–1840
Zhu W, Miao J, Qing L (2015) Constrained extreme learning machines: A study on classification cases. Computer Science
Hong S, Xiao C, Ma T, Li H, Sun J (2019) Mina: Multilevel knowledge-guided attention for modeling electrocardiography signals. In: Twenty-eighth international joint conference on artificial intelligence IJCAI-19
Verma KK, Singh BM (2021) Deep multi-model fusion for human activity recognition using evolutionary algorithms. International Journal of Interactive Multimedia & Artificial Intelligence, 7(2)
Kocev D, Vens C, Struyf J, Džeroski S (2013) Tree ensembles for predicting structured outputs. Pattern Recogn 46(3):817–833
Read J, Reutemann P, Pfahringer B, Holmes G (2016) Meka: a multi-label/multi-target extension to weka
Tsoumakas G, Vlahavas I (2007) Random k-labelsets: An ensemble method for multilabel classification. In: European conference on machine learning, Springer, pp 406–417
Zhang Q-W, Zhong Y, Zhang M-L (2018) Feature-induced labeling information enrichment for multi-label learning. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol 32
Wu X-Z, Zhou Z-H (2017) A unified view of multi-label performance measures. In: International Conference on Machine Learning, PMLR, pp 3780–3788
Wang S, Minku LL, Ghezzi D, Caltabiano D, Tino P, Yao X (2013) Concept drift detection for online class imbalance learning. In: The 2013 International Joint Conference on Neural Networks (IJCNN), IEEE, pp 1–10
Katakis I, Tsoumakas G, Vlahavas IP (2008) An ensemble of classifiers for coping with recurring contexts in data streams.. In: ECAI, pp 763–764
Xia Y, Chen K, Yang Y (2021) Multi-label classification with weighted classifier selection and stacked ensemble. Inf Sci 557:421–442
Li P, Wang H, Böhm C, Shao J (2020) Online semi-supervised multi-label classification with label compression and local smooth regression.. In: IJCAI, pp 1359–1365
Acknowledgements
This work was supported in part by the Hebei Natural Science Foundation No.G2021203010 and No.F2021203038 in China, and a project supported by Key Laboratory of Robotics and Intelligent Equipment of Guangdong Regular Institutions of Higher EducationGrant No.2017KSYS009.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
This article belongs to the Topical Collection: Big Data-Driven Large-Scale Group Decision Making Under Uncertainty
Rights and permissions
About this article
Cite this article
Liang, S., Pan, W., You, D. et al. Incremental deep forest for multi-label data streams learning. Appl Intell 52, 13398–13414 (2022). https://doi.org/10.1007/s10489-022-03414-6
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-022-03414-6