Abstract
Distant supervision (DS) has the advantage of automatically annotate large amounts of data and has been widely used for relation classification. Despite its efficiency, it often suffers from the label noise problem, which would impair the performance of relation classification. Recently, there are two ways to solve the label noise problem. The first way is to use multi-instance learning to consider the noises of instances, but they do not perform well for sentence-level prediction. The second way is to use reinforcement learning or adversarial learning to directly find noisy label instances but with high computational overhead and poor performance. In this paper, we propose to use the natural language inference (NLI) model to evaluate the quality of the instances directly, and select the high-quality instances as refined training data for sentence-level relation classification. Due to the lack of high-quality supervised data, we use reinforcement learning to train the NLI model. Experimental results on two human re-annotated NYT datasets show the effectiveness and efficiency of our method at the sentence-level relation classification. The source code of this paper can be found in https://github.com/xubodhu/RLRC.
This paper was supported by the National Natural Science Foundation of China (61906035), Shanghai Sailing Program (19YF1402300) and National Natural Science Foundation of China (61972081).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Feng, J., Huang, M., Zhao, L., Yang, Y., Zhu, X.: Reinforcement learning for relation classification from noisy data. In: Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, pp. 5779–5786 (2018)
Hoffmann, R., Zhang, C., Ling, X., Zettlemoyer, L., Weld, D.S.: Knowledge-based weak supervision for information extraction of overlapping relations. In: Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, pp. 541–550 (2011)
Lin, Y., Shen, S., Liu, Z., Luan, H., Sun, M.: Neural relation extraction with selective attention over instances. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, pp. 2124–2133 (2016)
Mintz, M., Bills, S., Snow, R., Jurafsky, D.: Distant supervision for relation extraction without labeled data. In: Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL, pp. 1003–1011. Association for Computational Linguistics (2009)
Phi, V.T., Santoso, J., Tran, V.H., Shindo, H., Shimbo, M., Matsumoto, Y.: Distant supervision for relation extraction via piecewise attention and bag-level contextual inference. IEEE Access 7, 103570–103582 (2019)
Qin, P., Xu, W., Wang, W.Y.: DSGAN: generative adversarial training for distant supervision relation extraction. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, pp. 496–505 (2018)
Qin, P., Xu, W., Wang, W.Y.: Robust distant supervision relation extraction via deep reinforcement learning. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, pp. 2137–2147 (2018)
Riedel, S., Yao, L., McCallum, A.: Modeling relations and their mentions without labeled text. In: Balcázar, J.L., Bonchi, F., Gionis, A., Sebag, M. (eds.) ECML PKDD 2010. LNCS (LNAI), vol. 6323, pp. 148–163. Springer, Heidelberg (2010). https://doi.org/10.1007/978-3-642-15939-8_10
Su, P., Li, G., Wu, C., Vijay-Shanker, K.: Using distant supervision to augment manually annotated data for relation extraction. PloS one 14(7), 1–17 (2019)
Surdeanu, M., Tibshirani, J., Nallapati, R., Manning, C.D.: Multi-instance multi-label learning for relation extraction. In: Proceedings of the 2012 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning, pp. 455–465 (2012)
Williams, R.J.: Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine Learn. 8(3–4), 229–256 (1992)
Wu, Y., Bamman, D., Russell, S.: Adversarial training for relation extraction. In: Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pp. 1778–1783 (2017)
Yoon, J., Arik, S., Pfister, T.: Data valuation using reinforcement learning. In: International Conference on Machine Learning, pp. 10842–10851. PMLR (2020)
Zeng, D., Liu, K., Chen, Y., Zhao, J.: Distant supervision for relation extraction via piecewise convolutional neural networks. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pp. 1753–1762 (2015)
Zeng, D., Liu, K., Lai, S., Zhou, G., Zhao, J.: Relation classification via convolutional deep neural network. In: Proceedings of the 25th International Conference on Computational Linguistics: Technical Papers, pp. 2335–2344 (2014)
Zeng, X., He, S., Liu, K., Zhao, J.: Large scaled relation extraction with reinforcement learning. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 5658–5665 (2018)
Zhang, A., Lipton, Z.C., Li, M., Smola, A.J.: Dive into Deep Learning. Corwin (2020)
Zhu, T., Wang, H., Yu, J., Zhou, X., Chen, W., Zhang, W., Zhang, M.: Towards accurate and consistent evaluation: a dataset for distantly-supervised relation extraction. In: Proceedings of the 28th International Conference on Computational Linguistics, pp. 6436–6447 (2020)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Xu, B., Zhao, X., Sha, C., Zhang, M., Song, H. (2021). Reinforced Natural Language Inference for Distantly Supervised Relation Classification. In: Karlapalem, K., et al. Advances in Knowledge Discovery and Data Mining. PAKDD 2021. Lecture Notes in Computer Science(), vol 12714. Springer, Cham. https://doi.org/10.1007/978-3-030-75768-7_29
Download citation
DOI: https://doi.org/10.1007/978-3-030-75768-7_29
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-75767-0
Online ISBN: 978-3-030-75768-7
eBook Packages: Computer ScienceComputer Science (R0)