Skip to main content

Utilizing BERT Pretrained Models with Various Fine-Tune Methods for Subjectivity Detection

  • Conference paper
  • First Online:
Book cover Web and Big Data (APWeb-WAIM 2020)

Abstract

As an essentially antecedent task of sentiment analysis, subjectivity detection refers to classifying sentences to be subjective ones containing opinions, or objective and neutral ones without bias. In the situations where impartial language is required, such as Wikipedia, subjectivity detection could play an important part. Recently, pretrained language models have proven to be effective in learning representations, profoundly boosting the performance among several NLP tasks. As a state-of-art pretrained model, BERT is trained on large unlabeled data with masked word prediction and next sentence prediction tasks. In this paper, we mainly explore utilizing BERT pretrained models with several combinations of fine-tuning methods, holding the intention to enhance performance in subjectivity detection task. Our experimental results reveal that optimum combinations of fine-tune and multi-task learning surplus the state-of-the-art on subjectivity detection and related tasks.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Aleksandrova, D., Lareau, F., Ménard, P.: Multilingual Sentence-Level Bias Detection in Wikipedia (2019)

    Google Scholar 

  2. Burges, C.J.C., Ragno, R., Le, Q.V.: Learning to rank with nonsmooth cost functions. In: NIPS (2006)

    Google Scholar 

  3. Caruana, R.: Multitask learning: A knowledge-based source of inductive bias. Mach. Learn. 28, 41–75 (1997)

    Article  Google Scholar 

  4. Cer, D., Diab, M., Agirre, E., Lopez-Gazpio, I., Specia, L.: SemEval2017 Task 1: Semantic textual similarity multilingual and crosslingual focused evaluation. In: Proceedings f the 11th International Workshop Semantic Evaluation, August 2017, pp. 1–14

    Google Scholar 

  5. Devlin, J., Chang, M.-W., Lee, K., Toutanova, K.: Bert: pre-training of deep bidirectional transformers for language understanding (2018)

    Google Scholar 

  6. Hube, C., Fetahu, B.: Detecting biased statements in Wikipedia. In: The Web Conference, 1779–1786. International World Wide Web Conferences Steering Committee (2018)

    Google Scholar 

  7. Howard, J., Ruder, S.: Universal language model fine-tuning for text classification. In: Proceedings of the 56th Annual Meeting Association Computing Linguistics, vol. 1, pp. 328–339 (2018)

    Google Scholar 

  8. Hube, C., Fetahu, B.: Neural based statement classification for biased language. In: Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining, 195–203. ACM (2019)

    Google Scholar 

  9. Karamibekr, M., Ghorbani, A.A.: Sentence subjectivity analysis in social domains. In: Proceedings of the IEEE/WIC/ACM International Conference on Web Intelligence (WI 2013), Atlanta, GA, USA (2013)

    Google Scholar 

  10. Kingma, D., Ba, J.: Adam: a method for stochastic optimization. In: ICLR (2015)

    Google Scholar 

  11. Lin, C., He, Y., Everson, R.: Sentence subjectivity detection with weakly-supervised learning. In: Proceedings of the 5th International Joint Conference on Natural Language Process, pp. 1153–1161, November 2011

    Google Scholar 

  12. Liu, X., He, P., Chen, W., Gao, J.: Multi-task deep neural networks for natural language understanding. In: Proceedings of the 57th Annual Meeting Association Computing Linguistics, pp. 4487–4496 (2019)

    Google Scholar 

  13. Maas, A.L., Daly, R.E., Pham, P.T., Huang, D., Ng, A.Y., Potts, C.: Learning word vectors for sentiment analysis. In: Proceedings of the 49th Annual Meeting Association Computing Linguistics, Human Language Technology vol. 1, pp. 142–150 (2011)

    Google Scholar 

  14. Montoyo, A., Martínez-Barco, P., Balahur, A.: Subjectivity and sentiment analysis: an overview of the current state of the area and envisaged developments. Decis. Supp. Syst. 53(4), 675–679 (2012)

    Google Scholar 

  15. Mäntylä, M.V., Graziotin, D., Kuutila, M.: The evolution of sentiment analysis—a review of research topics venues and top cited papers. Comput. Sci. Rev. 27, 16–32 (2018)

    Article  Google Scholar 

  16. Pang, B., Lee, L.: A sentimental education: Sentiment analysis using subjectivity summarization based on minimum cuts. In: Proceedings of ACL 2004 (2004)

    Google Scholar 

  17. Recasens, M., Danescu-Niculescu-Mizil, C., Jurafsky, D.: Linguistic models for analyzing and detecting biased language. In: Proceedings of ACL (2013)

    Google Scholar 

  18. Rajpurkar, P., Zhang, J., Lopyrev, K., Liang, P.: Squad: 100,000 + Questions for Machine Comprehension of Text. In: EMNLP (2016)

    Google Scholar 

  19. Ruder, S.: An overview of multi-task learning in deep neural networks (2017). http://arxiv.org/abs/1706.05098

  20. Radford, A., Narasimhan, K., Salimans, T., Sutskever, I.: Improving language understanding by generative pre-training (2018). https://blog.openai.com/language-unsupervised/

  21. Bowman, S.R., Angeli, G., Potts, C., Manning, C,D.: A large annotated corpus for learning natural language inference. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, EMNLP 2015, Lisbon, Portugal, 17–21 September 2015, pp. 632–642 (2015)

    Google Scholar 

  22. Smith, L.N.: Cyclical learning rates for training neural networks. In: WACV (2017)

    Google Scholar 

  23. Smith, L.N.: A disciplined approach to neural network hyper-parameters: Part 1–learning rate, batch size, momentum, and weight decay. arXiv preprint arXiv:1803.09820 (2018)

  24. Sun, C., Qiu, X., Xu, Y., Huang, X.: How to fine-tune BERT for text classification? arXiv:1905.05583 (2019). https://arxiv.org/abs/1905.05583

  25. A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, I. Polosukhin, “Attention Is All You Need”, CoRR, vol. abs/1706.03762, 2017

    Google Scholar 

  26. Yang, Z., Dai, Z., Yang, Y., Carbonell, J., Salakhutdinov, R., Le, Q.V.: XLNet: generalized autoregressive pretraining for language understanding arXiv:1906.08237 (2019). https://arxiv.org/abs/1906.08237

  27. Zhao, H., Lu, Z., Poupart, P.: Self-adaptive hierarchical sentence model (2015). https://arxiv.org/abs/1504.05070

  28. Zhang, Y., Yang, Q.: A survey on multi-task learning. arXiv preprint arXiv:1707.08114 (2017)

  29. Zhang, L., Wang, S., Liu, B.: Deep learning for sentiment analysis: a survey. In: Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, p. 25, March 2018

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to Hairong Huo or Mizuho Iwaihara .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Huo, H., Iwaihara, M. (2020). Utilizing BERT Pretrained Models with Various Fine-Tune Methods for Subjectivity Detection. In: Wang, X., Zhang, R., Lee, YK., Sun, L., Moon, YS. (eds) Web and Big Data. APWeb-WAIM 2020. Lecture Notes in Computer Science(), vol 12318. Springer, Cham. https://doi.org/10.1007/978-3-030-60290-1_21

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-60290-1_21

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-60289-5

  • Online ISBN: 978-3-030-60290-1

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics