Skip to main content

Prompting Generative Language Model with Guiding Augmentation for Aspect Sentiment Triplet Extraction

  • Conference paper
  • First Online:
PRICAI 2023: Trends in Artificial Intelligence (PRICAI 2023)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 14326))

Included in the following conference series:

  • 482 Accesses

Abstract

Aspect Sentiment Triplet Extraction (ASTE) is a challenging task in natural language processing concerning the automatic extraction of (aspect term, opinion term, sentiment polarity) triplets from a given text. Current end-to-end generative methods achieved high results by treating it as a sequence generation task with a generative pretrained language mode (e.g., T5). However, these architectures usually suffered from the objective gaps between the pre-training tasks and fine-tuning tasks, leading to suboptimal results. Further more, they can only provide information on what is a valid triplet, but no explicit guidance on what is not a triplet, which can not fully capture the correlation between aspects and opinions. To address above issues, we propose the generative prompt to bridge the gap between pre-training and fine-tuning of generative pretrained language model via text infilling task. And we propose guiding augmentation, which drops the aspect or opinion in the sentence by depicting a tree structure to generate diverse similar sentences and new target sequences. In this way, the main differences between these augmented samples are the dropped aspect or opinion term, and the model can understand the ASTE task knowledge better through the explicit variant constraints. Experimental results confirm that our method outperforms previous state-of-the-art (SOTA) methods on four public ASTE datasets.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 89.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 119.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Chen, H., Zhai, Z., Feng, F., Li, R., Wang, X.: Enhanced multi-channel graph convolutional network for aspect sentiment triplet extraction. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (vol. 1: Long Papers), pp. 2974–2985 (2022)

    Google Scholar 

  2. Chen, S., Wang, Y., Liu, J., Wang, Y.: Bidirectional machine reading comprehension for aspect sentiment triplet extraction. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 12666–12674 (2021)

    Google Scholar 

  3. Jing, H., Li, Z., Zhao, H., et al.: Seeking common but distinguishing difference, a joint aspect-based sentiment analysis model. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 3910–3922 (2021)

    Google Scholar 

  4. Lewis, M., Liu, Y., et al.: BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 7871–7880 (2020)

    Google Scholar 

  5. Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 (2017)

  6. Mao, Y., Shen, Y., Yang, J., Zhu, X., Cai, L.: Seq2Path: generating sentiment tuples as paths of a tree. In: Findings of the Association for Computational Linguistics: ACL 2022, pp. 2215–2225 (2022)

    Google Scholar 

  7. Peng, H., Xu, L., Bing, L., Huang, F., Lu, W., Si, L.: Knowing what, how and why: a near complete solution for aspect-based sentiment analysis. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, pp. 8600–8607 (2020)

    Google Scholar 

  8. Petroni, F., Rocktäschel, T., Riedel, S., et al.: Language models as knowledge bases? In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 2463–2473 (2019)

    Google Scholar 

  9. Pontiki, M., Galanis, D., et al.: SemEval-2015 task 12: aspect based sentiment analysis. In: Proceedings of the 9th International Workshop on Semantic evaluation (SemEval 2015), pp. 486–495 (2015)

    Google Scholar 

  10. Pontiki, M., et al.: SemEval-2016 task 5: aspect based sentiment analysis. In: ProWorkshop on Semantic Evaluation (SemEval-2016), pp. 19–30. Association for Computational Linguistics (2016)

    Google Scholar 

  11. Pontiki, M., Papageorgiou, H., et al.: SemEval-2014 task 4: aspect based sentiment analysis. In: SemEval 2014, p. 27 (2014)

    Google Scholar 

  12. Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21, 5485–5551 (2020)

    MathSciNet  Google Scholar 

  13. Wu, Z., Ying, C., Zhao, F., Fan, Z., Dai, X., Xia, R.: Grid tagging scheme for aspect-oriented fine-grained opinion extraction. In: Findings of the Association for Computational Linguistics: EMNLP 2020, pp. 2576–2585 (2020)

    Google Scholar 

  14. Xu, L., Chia, Y.K., Bing, L.: Learning span-level interactions for aspect sentiment triplet extraction. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (vol. 1: Long Papers), pp. 4755–4766 (2021)

    Google Scholar 

  15. Zhai, Z., Chen, H., Feng, F., Li, R., Wang, X.: COM-MRC: a context-masked machine reading comprehension framework for aspect sentiment triplet extraction. In: Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pp. 3230–3241 (2022)

    Google Scholar 

  16. Zhang, W., Deng, Y., Li, X., Yuan, Y., Bing, L., Lam, W.: Aspect sentiment quad prediction as paraphrase generation. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 9209–9219 (2021)

    Google Scholar 

  17. Zhang, W., Li, X., Deng, Y., Bing, L., Lam, W.: Towards generative aspect-based sentiment analysis. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (vol. 2: Short Papers), pp. 504–510 (2021)

    Google Scholar 

  18. Zhang, Y., et al.: Boundary-driven table-filling for aspect sentiment triplet extraction. In: Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pp. 6485–6498 (2022)

    Google Scholar 

Download references

Acknowledgement

This work was supported by Strategic Priority Research Program of Chinese Academy of Sciences (No. XDC02040400).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yongxiu Xu .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Huang, K., Xu, Y., Zhang, X., Zhang, W., Xu, H. (2024). Prompting Generative Language Model with Guiding Augmentation for Aspect Sentiment Triplet Extraction. In: Liu, F., Sadanandan, A.A., Pham, D.N., Mursanto, P., Lukose, D. (eds) PRICAI 2023: Trends in Artificial Intelligence. PRICAI 2023. Lecture Notes in Computer Science(), vol 14326. Springer, Singapore. https://doi.org/10.1007/978-981-99-7022-3_22

Download citation

  • DOI: https://doi.org/10.1007/978-981-99-7022-3_22

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-99-7021-6

  • Online ISBN: 978-981-99-7022-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics