Abstract
Current distant supervised relation extraction algorithms based on neural network deep learning always label a large number of irrelevant sentences as valid data or sentences as wrong relation because of their overly broad assumptions. Although many scholars have proposed some methods to reduce the influence of distant supervision noise and their optimization methods, such as organizing sentences into bag of sentences and introducing attention mechanism, they still cannot obtain a good feature vector of a bag of sentences. In this paper, we propose an attention-weight allocation algorithm for sentences inside a bag of sentences based on relative alignment. Compared with previous methods, it can better capture the similarity of sentences to relative vectors, and better extract relationships in sentences under distant supervision. On the standard dataset, the proposed model has improved by 2.5%, 2.4%, 2.8% and 2% compared with the PCNN_ATT model, P@100, P@200, P@300 and AUC indexes, respectively.
Keywords
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Zeng, D., Liu, K., Chen, Y., Zhao, J.: Distant supervision for relation extraction via piecewise convolutional neural networks. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pp. 1753–1762 (2015)
Mintz, M., Bills, S., Snow, R., Jurafsky, D.: Distant supervision for relation extraction without labeled data. In: Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP, pp. 1003–1011 (2009)
Riedel, S., Yao, L., McCallum, A.: Modeling relations and their mentions without labeled text. In: Balcázar, J.L., Bonchi, F., Gionis, A., Sebag, M. (eds.) ECML PKDD 2010. LNCS (LNAI), vol. 6323, pp. 148–163. Springer, Heidelberg (2010). https://doi.org/10.1007/978-3-642-15939-8_10
Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473 (2014)
Lin, Y., Shen, S., Liu, Z.: Neural relation extraction with selective attention over instances. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, vol. 1, pp. 2124–2133 (2016)
Ji, G., Liu, K., He, S., Zhao, J.: Distant supervision for relation extraction with sentence-level attention and entity descriptions. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 31, no. 1, pp. 3060–3066 (2017)
Mikolov, T., Sutskever, I., Chen, K., Corrado, G.S., Dean, J.: Distributed representations of words and phrases and their compositionality. In: Advances in neural information processing systems, pp. 3111–3119 (2013)
Vaswani, A., Shazeer, N., Parmar, N.: Attention is all you. In: Advances in neural information processing systems, pp. 5998–6008 (2017)
Du, J., Han, J., Way, A., Wan, D.: Multi-level structured self-attentions for distantly supervised relation extraction. arXiv preprint arXiv:1809.00699 (2018)
Yue, Q., Li, X., Li, D.: Chinese relation extraction on forestry knowledge graph construction. Compu. Syst. Sci. Eng. 37(3), 423–442 (2021)
Ding, K., Liu, S., Zhang, Y., Zhang, H., Zhang, X.: A knowledge-enriched and span-based network for joint entity and relation extraction. Comput. Mater. Continu. 68(1), 377–389 (2021)
Sun, J., Li, Y., Shen, Y., Ding, W., Shi, X.: Joint self-attention based neural networks for semantic relation extraction. J. Inform. Hiding Privacy Protect. 1(2), 69–75 (2019)
Zhao, S., Hu, M., Cai, Z., Z, Liu, F.: Dynamic modeling cross-modal interactions in two- phaseprediction for entity-relation extraction. IEEE Trans. Neural Netw. Learn. Syst. (2021) https://doi.org/10.1109/TNNLS.2021.3104971
Acknowledgement
The authors would like to thank the anonymous referees for their valuable comments and helpful suggestions. The work is supported by Science and Technology Project of the State Grid Information & Telecommunication Branch, “The research and technology for multi-Scenario Oriented security protection framework” (52993920002J).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Li, J., Huang, X., Gao, Y., Liu, J., Zhang, R., Zhao, J. (2022). Distant Supervised Relation Extraction Based on Sentence-Level Attention with Relation Alignment. In: Sun, X., Zhang, X., Xia, Z., Bertino, E. (eds) Artificial Intelligence and Security. ICAIS 2022. Lecture Notes in Computer Science, vol 13338. Springer, Cham. https://doi.org/10.1007/978-3-031-06794-5_12
Download citation
DOI: https://doi.org/10.1007/978-3-031-06794-5_12
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-06793-8
Online ISBN: 978-3-031-06794-5
eBook Packages: Computer ScienceComputer Science (R0)