Abstract
Argument pair extraction (APE) is an important research area for real-world applications such as online debates and persuasive essays. To facilitate the research of APE, NLPCC 2021 shared Task1 releases the RR-dataset, which aims to extract argument pairs from the peer reviews and its rebuttals. We propose a two-stage learning strategy for this task which divides it into two sub-tasks: arguments mining and arguments pairing. In arguments pairing task, instead of modeling arguments pairing task in sentence level, we cast it as a paragraph-level pairing task, which can alleviate the mismatch between the definition of task and training. And we apply transfer learning and fine-tuning strategy on all sub-tasks, which exploits large scale pre-trained semantic knowledge to benefit downstream APE task. Experiment results show that our method achieves significant improvements compared with strong baseline BERT-based multi-task learning framework, and finally ranks the \(3^{rd}\) in NLPCC2021 shared task 1 track 3 evaluation phrase.
S. Wang—Author contributed equally.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Boltužić, F., Šnajder, J.: Identifying prominent arguments in online debates using semantic textual similarity. In: Proceedings of the 2nd Workshop on Argumentation Mining, pp. 110–115 (2015)
Chakrabarty, T., Hidey, C., Muresan, S., McKeown, K., Hwang, A.: Ampersand: argument mining for persuasive online discussions. arXiv preprint arXiv:2004.14677 (2020)
Cheng, L., Bing, L., Yu, Q., Lu, W., Si, L.: Argument pair extraction from peer review and rebuttal via multi-task learning. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 7000–7011 (2020)
Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: Bert: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)
Dong, L., et al.: Unified language model pre-training for natural language understanding and generation. arXiv preprint arXiv:1905.03197 (2019)
Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)
Hou, Y., Jochim, C.: Argument relation classification using a joint inference model. In: Proceedings of the 4th Workshop on Argument Mining, pp. 60–66 (2017)
Hua, X., Nikolov, M., Badugu, N., Wang, L.: Argument mining for understanding peer reviews. arXiv preprint arXiv:1903.10104 (2019)
Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)
Lawrence, J., Reed, C.: Argument mining: a survey. Comput. Linguist. 45(4), 765–818 (2020)
Liu, Y., et al.: Roberta: a robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 (2019)
Müller, R., Kornblith, S., Hinton, G.: When does label smoothing help? arXiv preprint arXiv:1906.02629 (2019)
Persing, I., Ng, V.: End-to-end argumentation mining in student essays. In: Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 1384–1394 (2016)
Rocha, G., Stab, C., Cardoso, H.L., Gurevych, I.: Cross-lingual argumentative relation identification: from English to Portuguese. In: Proceedings of the 5th Workshop on Argument Mining, pp. 144–154 (2018)
Shnarch, E., et al.: Will it blend? blending weak and strong labeled data in a neural network for argumentation mining. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, vol. 2: Short Papers, pp. 599–605 (2018)
Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., Wojna, Z.: Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2818–2826 (2016)
Trautmann, D., Daxenberger, J., Stab, C., Schütze, H., Gurevych, I.: Fine-grained argument unit recognition and classification. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, pp. 9048–9056 (2020)
Wolf, T., et al.: Transformers: state-of-the-art natural language processing. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pp. 38–45. Association for Computational Linguistics, Online (2020). https://www.aclweb.org/anthology/2020.emnlp-demos.6
Xiong, W., Litman, D.: Automatically predicting peer-review helpfulness. In: Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, pp. 502–507 (2011)
Yang, Z., Dai, Z., Yang, Y., Carbonell, J., Salakhutdinov, R.R., Le, Q.V.: Xlnet: generalized autoregressive pretraining for language understanding. In: Advances in Neural Information Processing Systems, vol. 32 (2019)
Acknowledgements
We thank the reviewers for their careful reviewing and valuable advises, which are important for us to improve our work.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Wang, S., Yin, Z., Zhang, W., Zheng, D., Li, X. (2021). Two Stage Learning for Argument Pairs Extraction. In: Wang, L., Feng, Y., Hong, Y., He, R. (eds) Natural Language Processing and Chinese Computing. NLPCC 2021. Lecture Notes in Computer Science(), vol 13029. Springer, Cham. https://doi.org/10.1007/978-3-030-88483-3_44
Download citation
DOI: https://doi.org/10.1007/978-3-030-88483-3_44
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-88482-6
Online ISBN: 978-3-030-88483-3
eBook Packages: Computer ScienceComputer Science (R0)