Skip to main content

Investigating Active Positive-Unlabeled Learning with Deep Networks

  • Conference paper
  • First Online:
AI 2021: Advances in Artificial Intelligence (AI 2022)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 13151))

Included in the following conference series:

Abstract

Positive-unlabeled (PU) learning deals with the binary classification problem when only positive (P) and unlabeled (U) data are available. Recently, many PU learning models have been proposed based on deep networks and become the SOTA of PU learning. Despite the achievements on the model aspect, theoretical analysis and empirical results have shown that the number and quality of positive data can significantly impact learning performance. Active learning is classically used in machine learning to acquire additional high-quality labelled data, however, there are only a few studies on using active learning in deep PU models. This paper investigates the use of active learning in deep PU models. Specifically, this paper studies the uncertainty query strategy for pool-based active learning and show that due to the “large-small-loss” property of deep networks, the query strategy based purely on uncertainty can achieve diversity simultaneously. Empirical results also illustrate the effectiveness of uncertainty-based queries on active PU learning with deep networks.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 89.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 119.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. A survey of deep active learning. arXiv:2009.00236 (2020)

  2. Arpit, D., et al.: A closer look at memorization in deep networks. In: ICML (2017)

    Google Scholar 

  3. Chen, X., et al.: Self-PU: self boosted and calibrated positive-unlabeled training. In: ICML, pp. 1510–1519 (2020)

    Google Scholar 

  4. Denis, F.Ç.: PAC learning from positive statistical queries. In: Richter, M.M., Smith, C.H., Wiehagen, R., Zeugmann, T. (eds.) ALT 1998. LNCS (LNAI), vol. 1501, pp. 112–126. Springer, Heidelberg (1998). https://doi.org/10.1007/3-540-49730-7_9

    Chapter  Google Scholar 

  5. Du Plessis, M., Niu, G., Sugiyama, M.: Convex formulation for learning from positive and unlabeled data. In: ICML, pp. 1386–1394 (2015)

    Google Scholar 

  6. Ghasemi, A., Rabiee, H.R., Fadaee, M., Manzuri, M.T., Rohban, M.H.: Active learning from positive and unlabeled data. In: ICDM Workshop, pp. 244–250 (2011)

    Google Scholar 

  7. Kiryo, R., Niu, G., du Plessis, M.C., Sugiyama, M.: Positive-unlabeled learning with non-negative risk estimator. In: NIPS (2017)

    Google Scholar 

  8. Lee, W.S., Liu, B.: Learning with positive and unlabeled examples using weighted logistic regression. In: ICML, pp. 448–455 (2003)

    Google Scholar 

  9. Li, X., Liu, B.: Learning to classify texts using positive and unlabeled data. In: IJCAI, pp. 587–592 (2003)

    Google Scholar 

  10. Liu, B., Dai, Y., Li, X., Lee, W.S., Yu, P.S.: Building text classifiers using positive and unlabeled examples. In: ICDM, pp. 179–186 (2003)

    Google Scholar 

  11. Liu, B., Lee, W.S., Yu, P.S., Li, X.: Partially supervised classification of text documents. In: ICML, pp. 387–394 (2002)

    Google Scholar 

  12. Van der Maaten, L., Hinton, G.: Visualizing data using t-SNE. JMLR 9(11) (2008)

    Google Scholar 

  13. Niu, G., du Plessis, M.C., Sakai, T., Ma, Y., Sugiyama, M.: Theoretical comparisons of positive-unlabeled learning against positive-negative learning. In: NIPS, pp. 1199–1207 (2016)

    Google Scholar 

  14. Sakai, T., du Plessis, M.C., Niu, G., Sugiyama, M.: Semi-supervised classification based on classification from positive and unlabeled data. In: ICML, pp. 2998–3006 (2017)

    Google Scholar 

  15. Settles, B.: Active Learning. Morgan & Claypool Publishers, San Rafael (2012)

    Google Scholar 

  16. Su, G., Chen, W., Xu, M.: Positive-unlabeled learning from imbalanced data. In: IJCAI, pp. 2995–3001 (2021)

    Google Scholar 

  17. Xu, M., Li, B., Niu, G., Han, B., Sugiyama, M.: Revisiting sample selection approach to positive-unlabeled learning: turning unlabeled data into positive rather than negative. arXiv: 1901.10155 (2019)

Download references

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to Weitong Chen or Miao Xu .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Han, K., Chen, W., Xu, M. (2022). Investigating Active Positive-Unlabeled Learning with Deep Networks. In: Long, G., Yu, X., Wang, S. (eds) AI 2021: Advances in Artificial Intelligence. AI 2022. Lecture Notes in Computer Science(), vol 13151. Springer, Cham. https://doi.org/10.1007/978-3-030-97546-3_49

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-97546-3_49

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-97545-6

  • Online ISBN: 978-3-030-97546-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics