Abstract
Distant supervision, a method for relation extraction, leverages knowledge base triples to label entities and relations in text, but this leads to noisy labels and long-tail problems. Among long-tail dependency structures, the hierarchy tree of relations is the most classical and has demonstrated great efficacy in information extraction. However, the hierarchical tree of relations presents a challenge in obtaining sufficient information representation in cases where there is no sibling node or parent node without sibling node. To address this challenge, the use of constraint graphs has been proposed, but such approaches neglect the hierarchical information in the relations. To overcome this limitation, we propose a model based on dependency paths and relational graph embeddings. The model utilizes two relational graph structures, the constraint graph and the relation hierarchy tree, for relation learning, with the aim of transferring the knowledge learned in the data-rich relation to the long-tail relation. Additionally, the model leverages the shortest dependency path between entity pairs to increase the discriminative power of entity pairs in different bags for multi-instance learning. Experimental results show that the model achieves an AUC of 54.3% on the NYT-10 dataset and 86.3% on Hit@15 (<100).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Mintz, M., Bills, S., Snow, R., Jurafsky, D.: Distant supervision for relation extraction without labeled data. In: ACL-AFNLP, pp. 1003–1011 (2009)
Han, X., Yu, P., Liu, Z., Sun, M., Li, P.: Hierarchical relation extraction with coarse-to-fine grained attention. In: Proceedings of EMNLP, pp. 2236–2245 (2018)
Christou, D., Tsoumakas, G.: Improving distantly-supervised relation extraction through BERT-based label and instance embeddings. IEEE Access 9, 62574–62582 (2021)
Hoffmann, R., Zhang, C., Ling, X., Zettlemoyer, L., Weld, D.S.: Knowledge-based weak supervision for information extraction of overlapping relations. In: Proceedings of ACL HLT, pp. 541–550 (2011)
Yu, E., Han, W., Tian, Y., Chang, Y.: ToHRE: a top-down classification strategy with hierarchical bag representation for distantly supervised relation extraction. In: Proceedings of COLING, pp. 1665–1676 (2020)
Liang, T., Liu, Y., Liu, X., Zhang, H., Sharma, G., Guo, M.: Distantly-supervised long-tailed relation extraction using constraint graphs. IEEE Trans. Knowl. Data Eng. 35(7), 6852–6865 (2022)
Li, Y., et al.: Self-attention enhanced selective gate with entity-aware embedding for distantly supervised relation extraction. Proceedings of the AAAI 34(05), 8269–8276 (2020)
Han, X., Gao, T., Yao, Y., Ye, D., Liu, Z., Sun, M.: OpenNRE: An open and extensible toolkit for neural relation extraction (2019). arXiv preprint arXiv:1909.13078
Lin, Y., Shen, S., Liu, Z., Luan, H., Sun, M.: Neural relation extraction with selective attention over instances. In: Proceedings of ACL, pp. 2124–2133 (2016)
Vashishth, S., Joshi, R., Prayaga, S.S., Bhattacharyya, C., Talukdar, P.: RESIDE: Improving distantly-supervised neural relation extraction using side information (2018). arXiv preprint arXiv:1812.04361
Alt, C., Hübner, M., Hennig, L.: Fine-tuning pre-trained transformer language models to distantly supervised relation extraction (2019). arXiv preprint arXiv:1906.08646
Riedel, S., Yao, L., McCallum, A.: Modeling relations and their mentions without labeled text. In: Balcázar, J.L., Bonchi, F., Gionis, A., Sebag, M. (eds.) ECML PKDD 2010. LNCS (LNAI), vol. 6323, pp. 148–163. Springer, Heidelberg (2010). https://doi.org/10.1007/978-3-642-15939-8_10
Qu, J., Ouyang, D., Hua, W., Ye, Y., Li, X.: Distant supervision for neural relation extraction integrated with word attention and property features. Neural Netw. 100, 59–69 (2018)
Du, J., Han, J., Way, A., Wan, D.: Multi-level structured self-attentions for distantly supervised relation extraction (2018). arXiv preprint arXiv:1809.00699
Yuan, Y., et al.: Cross-relation cross-bag attention for distantly-supervised relation extraction. In: Proceedings of the AAAI, pp. 419–426 (2019)
Ye, Z.-X., Ling, Z.-H.: Distant supervision relation extraction with intra-bag and inter-bag attentions (2019). arXiv preprint arXiv:1904.00143
Dai, L., Xu, B., Song, H.: Feature-level attention based sentence encoding for neural relation extraction. In: Tang, J., Kan, M.-Y., Zhao, D., Li, S., Zan, H. (eds.) NLPCC 2019. LNCS (LNAI), vol. 11838, pp. 184–196. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-32233-5_15
Yu, B., Zhang, Z., Liu, T., Wang, B., Li, S., Li, Q.: Beyond word attention: using segment attention in neural relation extraction. In: IJCAI, pp. 5401–5407 (2019)
Gui, Y., Liu, Q., Zhu, M., Gao, Z.: Exploring long tail data in distantly supervised relation extraction. In: Lin, C.-Y., Xue, N., Zhao, D., Huang, X., Feng, Y. (eds.) ICCPOL/NLPCC -2016. LNCS (LNAI), vol. 10102, pp. 514–522. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-50496-4_44
Zhang, N., et al.: Long-tail relation extraction via knowledge graph embeddings and graph convolution networks (2019). arXiv preprint arXiv:1903.01306
Bordes, A., Usunier, N., Garcia-Duran, A., Weston, J., Yakhnenko, O.: Translating embeddings for modeling multi-relational data. In: Advances in Neural Information Processing Systems, vol. 26 (2013)
Defferrard, M., Bresson, X., Vandergheynst, P.: Convolutional neural networks on graphs with fast localized spectral filtering. In: Advances in Neural Information Processing Systems, vol. 29 (2016)
Cao, Y., Kuang, J., Gao, M., Zhou, A., Wen, Y., Chua, T.-S.: Learning relation prototype from unlabeled texts for long-tail relation extraction. IEEE Trans. Knowl. Data Eng. 35(2), 1761–1774 (2021)
Gou, Y., Lei, Y., Liu, L., Zhang, P., Peng, X.: A dynamic parameter enhanced network for distant supervised relation extraction. Knowl.-Based Syst. 197, 105912 (2020)
Zeng, D., Liu, K., Chen, Y., Zhao, J.: Distant supervision for relation extraction via piecewise convolutional neural networks. In: Proceedings of EMNLP, pp. 1753–1762 (2015)
Devlin, J., Chang, M.-W., Lee, K., Toutanova, K.: BERT: Pre-training of deep bidirectional transformers for language understanding (2018). arXiv preprint arXiv:1810.04805
Kipf, T.N., Welling, M.: Semi-supervised classification with graph convolutional networks (2016). arXiv preprint arXiv:1609.02907
Veličković, P., Cucurull, G., Casanova, A., Romero, A., Lio, P., Bengio, Y.: Graph attention networks (2017). arXiv preprint arXiv:1710.10903
Li, Y., Shen, T., Long, G., Jiang, J., Zhou, T., Zhang, C.: Improving long-tail relation extraction with collaborating relation-augmented attention (2020). arXiv preprint arXiv:2010.03773
Wang, J.: RH-Net: improving neural relation extraction via reinforcement learning and hierarchical relational searching (2020). arXiv preprint arXiv:2010.14255
Li, Y., Long, G., Shen, T., Jiang, J.: Hierarchical relation-guided type-sentence alignment for long-tail relation extraction with distant supervision (2021). arXiv preprint arXiv:2109.09036
Acknowledgements
This research was supported by the Dongbo Future Artificial Intelligence Research Institute Co., Ltd. Joint Laboratory (School Agreement No. 20223160C0026), Xiaozhi Deep Art Artificial Intelligence Research Institute Co., Ltd. Computational Art Joint Laboratory (School Agreement No. 20213160C0032), and Xiamen Yinjiang Smart City Joint Research Center (School Agreement No. 20213160C0029).
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Li, Y., Zong, Y., Sun, W., Wu, Q., Hong, Q. (2024). A Long-Tail Relation Extraction Model Based on Dependency Path and Relation Graph Embedding. In: Song, X., Feng, R., Chen, Y., Li, J., Min, G. (eds) Web and Big Data. APWeb-WAIM 2023. Lecture Notes in Computer Science, vol 14332. Springer, Singapore. https://doi.org/10.1007/978-981-97-2390-4_28
Download citation
DOI: https://doi.org/10.1007/978-981-97-2390-4_28
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-97-2389-8
Online ISBN: 978-981-97-2390-4
eBook Packages: Computer ScienceComputer Science (R0)