Abstract
Fine-grained sentiment analysis of dialogue text is crucial for the model to understand the conversational participants’ viewpoints and provide accurate responses in generating replies. Unfortunately, in the field of conversational opinion mining, coarse-grained dialogue emotion analysis remains the mainstream approach, despite being unable to meet the actual needs in some specific scenarios such as customer service question and answer system. This work focuses on conversational aspect-based sentiment quadruple analysis, which aims to detect the sentiment quadruple of target-aspect-opinion-sentiment in a dialogue. In this study, we mainly extract triplets and judge the unique sentiment, which is determined by the target and opinion terms together. For this purpose, we fine-tune the pre-trained language models using the DiaASQ dataset. We optimize the rotation positional information embedding by combining the actual length of the dialogue text and use adversarial training to enhance the model’s performance and robustness. Finally, We use beam search ensemble algorithm to improve the entire triplet extraction system’s performance. Our system achieved an average F1 score 40.50 that ranked second in the Chinese dataset and fifth in the general dataset for the Conversational Aspect-based Sentiment Quadruple Analysis shared task at NLPCC-2023.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Liu, B., Zhang, L.: A survey of opinion mining and sentiment analysis. In: Mining Text Data, pp. 415–463. Springer (2012)
Li, B., Fei, H.: Diaasq: a benchmark of conversational aspect-based sentiment quadruple analysis. CoRR abs/2211.05705 (2022)
Su, J., Lu, Y.: Roformer: enhanced transformer with rotary position embedding. CoRR abs/2104.09864 (2021)
Miyato, T., Dai, A.M.: Adversarial training methods for semi-supervised text classification. In: 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24–26, 2017, Conference Track Proceedings. OpenReview.net (2017)
Cui, S., Han, Y.: A two-stage voting-boosting technique for ensemble learning in social network sentiment classification. Entropy 25(4), 555 (2023)
Zhao, H., Huang, L.: Spanmlt: a span-based multi-task learning framework for pair-wise aspect and opinion terms extraction. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, Online, July 5–10, 2020, pp. 3239–3248 (2022)
Wu, S., Fei, H.: Learn from syntax: improving pair-wise aspect and opinion terms extraction with rich syntactic knowledge. In: Zhou, Z. (ed.) Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI 2021, Virtual Event / Montreal, Canada, 19–27 August 2021, pp. 3957–3963 (2021)
Peng, H., Xu, L.: Knowing what, how and why: a near complete solution for aspect based sentiment analysis. In: The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7–12, 2020, pp. 8600–8607. AAAI Press (2020)
Knoester, J., Frasincar, F.: Domain adversarial training for aspect-based sentiment analysis. In: Web Information Systems Engineering - WISE 2022–23rd International Conference, Biarritz, France, November 1–3, 2022, Proceedings. LNCS, vol. 13724, pp. 21–37 (2022)
Li, Z., Zou, Y.: Learning implicit sentiment in aspect-based sentiment analysis with supervised contrastive pre-training. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP 2021, Virtual Event/Punta Cana, Dominican Republic, 7–11 November, 2021, pp. 246–256. Association for Computational Linguistics (2021)
Devlin, J., Chang, M.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, Minneapolis, MN, USA, June 2–7, 2019, Volume 1 (Long and Short Papers), pp. 4171–4186 (2019)
Liu, Y., Ott, M.: Roberta: A robustly optimized BERT pretraining approach. CoRR abs/1907.11692 (2019)
Cui, Y., Yang, Z.: PERT: pre-training BERT with permuted language model. CoRR abs/2203.06906 (2022)
Cui, Y., Che, W.: Revisiting pre-trained models for Chinese natural language processing. In: Findings of the Association for Computational Linguistics: EMNLP 2020, Online Event, 16–20 November 2020. Findings of ACL, vol. EMNLP 2020, pp. 657–668
Barnes, J., Kurtz, R.: Structured sentiment analysis as dependency graph parsing. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, ACL/IJCNLP 2021, (Volume 1: Long Papers), Virtual Event, August 1–6, 2021
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Tu, Z., Zhang, B., Jiang, C., Wang, J., Lin, H. (2023). A Model Ensemble Approach for Conversational Quadruple Extraction. In: Liu, F., Duan, N., Xu, Q., Hong, Y. (eds) Natural Language Processing and Chinese Computing. NLPCC 2023. Lecture Notes in Computer Science(), vol 14304. Springer, Cham. https://doi.org/10.1007/978-3-031-44699-3_16
Download citation
DOI: https://doi.org/10.1007/978-3-031-44699-3_16
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-44698-6
Online ISBN: 978-3-031-44699-3
eBook Packages: Computer ScienceComputer Science (R0)