Skip to main content

Generating Fine-Grained Aspect Names from Movie Review Sentences Using Generative Language Model

  • Conference paper
  • First Online:
Information Integration and Web Intelligence (iiWAS 2023)

Abstract

This paper proposes a method for identifying an aspect highlighted in a sentence from a movie review, utilizing a generative language model. For example, the aspect “SFX Techniques” is identified for the sentence “The explosions in cosmic space were realistic.” Classically, aspects are commonly estimated in the field of opinion mining within product reviews with classification or extraction approaches. However, because the aspects of movie reviews are diverse and innumerable, they cannot be listed in advance. Thus, we propose a generation-based approach using a generative language model to identify the aspect of a review sentence. We adopt T5 (Text-to-Text Transfer Transformer), a modern generative language model, providing additional pre-training and fine-tuning to reduce the training data. To verify the effectiveness of the learning techniques thus adopted, we conducted an experiment incorporating reviews of Yahoo! movies. Manual labeling of the correctness and diversity of the aspect names generated shows that our method can generates a variety of fine-grained aspect names using little training data.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 69.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 89.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    Hugging Face Transformer: https://huggingface.co/docs/transformers/index.

  2. 2.

    sonoisa/t5-base-japanese: https://huggingface.co/sonoisa/t5-base-japanese.

  3. 3.

    SentenceTransformers: https://www.sbert.net/.

  4. 4.

    Hugging Face sentence-transformers https://huggingface.co/sentence-transformers/.

  5. 5.

    For the sake of translation and anonymization, the reviews are fictitious, as the experiment was in Japanese and uses real review sentences prepared by an individual.

References

  1. Angelidis, S., Lapata, M.: Summarizing opinions: aspect extraction meets sentiment prediction and they are both weakly supervised. In: Proceedings of the Conference on Empirical Methods in Natural Language Processing, pp. 3675–3686 (2018). https://doi.org/10.18653/v1/D18-1403

  2. Hasib, K.M., Towhid, N.A., Alam, M.G.R.: Online review based sentiment classification on bangladesh airline service using supervised learning. In: 5th International Conference on Electrical Engineering and Information Communication Technology (ICEEICT), pp. 1–6. IEEE (2021)

    Google Scholar 

  3. He, R., Lee, W.S., Ng, H.T., Dahlmeier, D.: An unsupervised neural attention model for aspect extraction. In: Proc. of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) (2017)

    Google Scholar 

  4. Jo, Y., Oh, A.H.: Aspect and sentiment unification model for online review analysis. In: Proc. of the Fourth ACM International Conference on Web Search and Data Mining, pp. 815–824 (2011). https://doi.org/10.1145/1935826.1935932

  5. Karimi, A., Rossi, L., Prati, A.: Adversarial training for aspect-based sentiment analysis with bert. In: 2020 25th International Conference on Pattern Recognition (ICPR), pp. 8797–8803 (2021). https://doi.org/10.1109/ICPR48806.2021.9412167

  6. Kim, R.Y.: Using online reviews for customer sentiment analysis. IEEE Eng. Manage. Rev. 49(4), 162–168 (2021)

    Article  Google Scholar 

  7. Lewis, M., et al.: BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics (2020). https://doi.org/10.18653/v1/2020.acl-main.703

  8. Liu, Y.: Fine-tune bert for extractive summarization. arXiv preprint arXiv:1903.10318 (2019)

  9. Peng, H., Xu, L., Bing, L., Huang, F., Lu, W., Si, L.: Knowing what, how and why: a near complete solution for aspect-based sentiment analysis. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34(05), pp. 8600–8607 (2020). https://doi.org/10.1609/aaai.v34i05.6383

  10. Pruksachatkun, Y., et al.: Intermediate-task transfer learning with pretrained models for natural language understanding: When and why does it work? In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 5231–5247 (2020)

    Google Scholar 

  11. Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21(140), 1–67 (2020)

    MathSciNet  Google Scholar 

  12. Ravi, K., Ravi, V.: A survey on opinion mining and sentiment analysis: tasks, approaches and applications. Knowl.-Based Syst. 89, 14–46 (2015). https://doi.org/10.1016/j.knosys.2015.06.015

    Article  Google Scholar 

  13. Rietzler, A., Stabinger, S., Opitz, P., Engl, S.: Adapt or get left behind: domain adaptation through BERT language model finetuning for aspect-target sentiment classification. In: Proceedings of the Twelfth Language Resources and Evaluation Conference, pp. 4933–4941 (2020)

    Google Scholar 

  14. Singh, V.K., Piryani, R., Uddin, A., Waila, P.: Sentiment analysis of movie reviews: a new feature-based heuristic for aspect-level sentiment classification. In: 2013 International Mutli-Conference on Automation, Computing, Communication, Control and Compressed Sensing (iMac4s), pp. 712–717 (2013). https://doi.org/10.1109/iMac4s.2013.6526500

  15. Xu, H., Liu, B., Shu, L., Yu, P.S.: Double embeddings and cnn-based sequence labeling for aspect extraction. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (2018)

    Google Scholar 

  16. Xu, H., Liu, B., Shu, L., Yu, P.S.: Bert post-training for review reading comprehension and aspect-based sentiment analysis. In: Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics (2019)

    Google Scholar 

  17. Zhang, J., Zhao, Y., Saleh, M., Liu, P.: PEGASUS: pre-training with extracted gap-sentences for abstractive summarization. In: International Conference on Machine Learning, pp. 11328–11339 (2020)

    Google Scholar 

Download references

Acknowledgements

This work was supported by JSPS KAKENHI Grants Number 21H03775, 21H03774, and 22H03905.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yoshiyuki Shoji .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Ishii, T., Shoji, Y., Yamamoto, T., Ohshima, H., Fujita, S., Dürst, M.J. (2023). Generating Fine-Grained Aspect Names from Movie Review Sentences Using Generative Language Model. In: Delir Haghighi, P., et al. Information Integration and Web Intelligence. iiWAS 2023. Lecture Notes in Computer Science, vol 14416. Springer, Cham. https://doi.org/10.1007/978-3-031-48316-5_23

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-48316-5_23

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-48315-8

  • Online ISBN: 978-3-031-48316-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics