Abstract
Positive-unlabeled (PU) learning deals with the binary classification problem when only positive (P) and unlabeled (U) data are available. Recently, many PU learning models have been proposed based on deep networks and become the SOTA of PU learning. Despite the achievements on the model aspect, theoretical analysis and empirical results have shown that the number and quality of positive data can significantly impact learning performance. Active learning is classically used in machine learning to acquire additional high-quality labelled data, however, there are only a few studies on using active learning in deep PU models. This paper investigates the use of active learning in deep PU models. Specifically, this paper studies the uncertainty query strategy for pool-based active learning and show that due to the “large-small-loss” property of deep networks, the query strategy based purely on uncertainty can achieve diversity simultaneously. Empirical results also illustrate the effectiveness of uncertainty-based queries on active PU learning with deep networks.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
A survey of deep active learning. arXiv:2009.00236 (2020)
Arpit, D., et al.: A closer look at memorization in deep networks. In: ICML (2017)
Chen, X., et al.: Self-PU: self boosted and calibrated positive-unlabeled training. In: ICML, pp. 1510–1519 (2020)
Denis, F.Ç.: PAC learning from positive statistical queries. In: Richter, M.M., Smith, C.H., Wiehagen, R., Zeugmann, T. (eds.) ALT 1998. LNCS (LNAI), vol. 1501, pp. 112–126. Springer, Heidelberg (1998). https://doi.org/10.1007/3-540-49730-7_9
Du Plessis, M., Niu, G., Sugiyama, M.: Convex formulation for learning from positive and unlabeled data. In: ICML, pp. 1386–1394 (2015)
Ghasemi, A., Rabiee, H.R., Fadaee, M., Manzuri, M.T., Rohban, M.H.: Active learning from positive and unlabeled data. In: ICDM Workshop, pp. 244–250 (2011)
Kiryo, R., Niu, G., du Plessis, M.C., Sugiyama, M.: Positive-unlabeled learning with non-negative risk estimator. In: NIPS (2017)
Lee, W.S., Liu, B.: Learning with positive and unlabeled examples using weighted logistic regression. In: ICML, pp. 448–455 (2003)
Li, X., Liu, B.: Learning to classify texts using positive and unlabeled data. In: IJCAI, pp. 587–592 (2003)
Liu, B., Dai, Y., Li, X., Lee, W.S., Yu, P.S.: Building text classifiers using positive and unlabeled examples. In: ICDM, pp. 179–186 (2003)
Liu, B., Lee, W.S., Yu, P.S., Li, X.: Partially supervised classification of text documents. In: ICML, pp. 387–394 (2002)
Van der Maaten, L., Hinton, G.: Visualizing data using t-SNE. JMLR 9(11) (2008)
Niu, G., du Plessis, M.C., Sakai, T., Ma, Y., Sugiyama, M.: Theoretical comparisons of positive-unlabeled learning against positive-negative learning. In: NIPS, pp. 1199–1207 (2016)
Sakai, T., du Plessis, M.C., Niu, G., Sugiyama, M.: Semi-supervised classification based on classification from positive and unlabeled data. In: ICML, pp. 2998–3006 (2017)
Settles, B.: Active Learning. Morgan & Claypool Publishers, San Rafael (2012)
Su, G., Chen, W., Xu, M.: Positive-unlabeled learning from imbalanced data. In: IJCAI, pp. 2995–3001 (2021)
Xu, M., Li, B., Niu, G., Han, B., Sugiyama, M.: Revisiting sample selection approach to positive-unlabeled learning: turning unlabeled data into positive rather than negative. arXiv: 1901.10155 (2019)
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 Springer Nature Switzerland AG
About this paper
Cite this paper
Han, K., Chen, W., Xu, M. (2022). Investigating Active Positive-Unlabeled Learning with Deep Networks. In: Long, G., Yu, X., Wang, S. (eds) AI 2021: Advances in Artificial Intelligence. AI 2022. Lecture Notes in Computer Science(), vol 13151. Springer, Cham. https://doi.org/10.1007/978-3-030-97546-3_49
Download citation
DOI: https://doi.org/10.1007/978-3-030-97546-3_49
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-97545-6
Online ISBN: 978-3-030-97546-3
eBook Packages: Computer ScienceComputer Science (R0)