Skip to main content
Log in

Dependency-position relation graph convolutional network with hierarchical attention mechanism for relation extraction

  • Published:
The Journal of Supercomputing Aims and scope Submit manuscript

Abstract

Existing research extensively incorporates syntactic information, especially dependency trees, to enhance the performance of relation extraction tasks. However, relying solely on dependency information may not fully exploit the rich semantic and syntactic information contained in sentences, and not all information in dependency trees is substantively helpful for relation extraction. Therefore, this paper proposes the Dependent-Position Relation Graph Convolutional network with Hierarchical Attention (DPR-GHA) for relation extraction, a method that integrates dependency relations and position relations into a hierarchical attention mechanism to effectively capture the relations between entities in text. The method aims to capture rich semantic information and enhance the performance of relation extraction. Specifically, we introduce the dependency relations of sentences and position relations of words to model global dependencies and local features, respectively. Subsequently, a novel hierarchical attention mechanism is introduced into the Graph Convolutional Network (GCN), dynamically adjusting the weights between nodes based on the input of the graph convolutional layer. This adaptive information aggregation enables each node to aggregate information adaptively according to its context and the importance of neighboring nodes. The research results on the SemEval-2010 Task 8 and KBP37 datasets thoroughly validate the effectiveness of the proposed model, demonstrating its significant performance advantage in relation extraction tasks.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12

Similar content being viewed by others

Data availability

No datasets were generated or analyzed during the current study.

References

  1. Yu M, Yin W, Hasan KS, Santos C, Xiang B, Zhou B (2017) Improved neural relation detection for knowledge base question answering. In: Proceedings of the 55th annual meeting of the association for computational linguistics (Volume 1: Long Papers), pp 571–581

  2. Kadry A, Dietz L (2017) Open relation extraction for support passage retrieval: merit and open issues. In: Proceedings of the 40th international ACM SIGIR conference on research and development in information retrieval, pp 1149–1152

  3. Zeng D, Liu K, Lai S, Zhou G, Zhao J (2014) Relation classification via convolutional deep neural network. In: Proceedings of COLING 2014, the 25th international conference on computational linguistics: technical papers, pp 2335–2344

  4. Zhou P, Shi W, Tian J, Qi Z, Li B, Hao H, Xu B (2016) Attention-based bidirectional long short-term memory networks for relation classification. In: Proceedings of the 54th annual meeting of the association for computational linguistics (vol. 2: Short Papers), pp 207–212

  5. Peng N, Poon H, Quirk C, Toutanova K, Yih WT (2017) Cross-sentence N-ary relation extraction with graph LSTMS. Trans Assoc Comput Linguist 5:101–115

    Article  Google Scholar 

  6. Sun K, Zhang R, Mao Y, Mensah S, Liu X (2020) Relation extraction with convolutional network over learnable syntax-transport graph. Proc AAAI Conf Artif Intell 34:8928–8935

    Google Scholar 

  7. Sun H, Grishman R (2022) Lexicalized dependency paths based supervised learning for relation extraction. Comput Syst Sci Eng 43(3)

  8. Hong Y, Liu Y, Yang S, Zhang K, Hu J (2020) Joint extraction of entities and relations using graph convolution over pruned dependency trees. Neurocomputing 411:302–312

    Article  Google Scholar 

  9. Jin Y, Li J, Lian Z, Jiao C, Hu X (2022) Supporting medical relation extraction via causality-pruned semantic dependency forest. In: Proceedings of the 29th international conference on computational linguistics, pp 2450–2460

  10. Chen T, Zhou L, Wang N, Chen X (2022) Joint entity and relation extraction with position-aware attention and relation embedding. Appl Soft Comput 119:108604

    Article  Google Scholar 

  11. Li Z, Sun Y, Tang S, Zhang C, Ma H (2021) Reinforcement learning with dual attention guided graph convolution for relation extraction. In: 2020 25th international conference on pattern recognition (ICPR), IEEE. pp 946–953

  12. Zelenko D, Aone C, Richardella A (2003) Kernel methods for relation extraction. J Mach Learn Res 3:1083–1106

    MathSciNet  Google Scholar 

  13. Bunescu R, Mooney R (2005) A shortest path dependency kernel for relation extraction. In: Proceedings of human language technology conference and conference on empirical methods in natural language processing, pp 724–731

  14. McDonald R, Pereira F, Kulick S, Winters S, Jin Y, White P (2005) Simple algorithms for complex relation extraction with applications to biomedical IE. In: Proceedings of the 43rd annual meeting of the association for computational linguistics (ACL’05), pp 491–498

  15. Mintz M, Bills S, Snow R, Jurafsky D (2009) Distant supervision for relation extraction without labeled data. In: Proceedings of the joint conference of the 47th annual meeting of the ACL and the 4th international joint conference on natural language processing of the AFNLP, pp 1003–1011

  16. Socher R, Huval B, Manning CD, Ng AY (2012) Semantic compositionality through recursive matrix-vector spaces. In: Proceedings of the 2012 joint conference on empirical methods in natural language processing and computational natural language learning, pp 1201–1211

  17. Vu NT, Adel H, Gupta P, Schütze H (2016) Combining recurrent and convolutional neural networks for relation classification. In: Proceedings of the 2016 conference of the North American chapter of the association for computational linguistics: human language technologies, pp 534–539

  18. Wu S, He Y (2019) Enriching pre-trained language model with entity information for relation classification. In: Proceedings of the 28th ACM international conference on information and knowledge management, pp 2361–2364

  19. Gori M, Monfardini G, Scarselli F (2005) A new model for learning in graph domains. In: Proceedings. 2005 IEEE international joint conference on neural networks, 2005, IEEE. vol. 2, pp 729–734.

  20. Dolgorsuren B, Xu W, Khan KU, Jeong B-S, Lee Y-K (2016) Sp2: spanner construction for shortest path computation on streaming graph. In: Proceedings of the sixth international conference on emerging databases: technologies, applications, and theory, pp 43–50

  21. Zhang Y, Qi P, Manning CD (2018) Graph convolution over pruned dependency trees improves relation extraction. In: Proceedings of the 2018 conference on empirical methods in natural language processing, pp 2205–2215

  22. Fu T-J, Li P-H, Ma W-Y (2019) GraphREL: modeling text as relational graphs for joint entity and relation extraction. In: Proceedings of the 57th annual meeting of the association for computational linguistics, pp 1409–1418

  23. Guo Z, Zhang Y, Lu W (2019) Attention guided graph convolutional networks for relation extraction. In: Proceedings of the 57th annual meeting of the association for computational linguistics, pp 241–251

  24. Shi Y, Xiao Y, Quan P, Lei M, Niu L (2021) Distant supervision relation extraction via adaptive dependency-path and additional knowledge graph supervision. Neural Netw 134:42–53

    Article  Google Scholar 

  25. Zheng H, Wen R, Chen X, Yang Y, Zhang Y, Zhang Z, Zhang N, Qin B, Ming X, Zheng Y (2021) PRGC: Potential relation and global correspondence based joint relational triple extraction. In: Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing (Vol. 1: Long Papers), pp 6225–6235

  26. Dong Y, Xu X (2023) Weighted-dependency with attention-based graph convolutional network for relation extraction. Neural Process Lett 55(9):12121–12142

    Article  Google Scholar 

  27. Mandya A, Bollegala D, Coenen F (2020) Graph convolution over multiple dependency sub-graphs for relation extraction. In: COLING, pp. 6424–6435. International committee on computational linguistics

  28. Manning CD, Surdeanu M, Bauer J, Finkel JR, Bethard S, McClosky D (2014) The stanford coreNLP natural language processing toolkit. In: Proceedings of 52nd annual meeting of the association for computational linguistics: system demonstrations, pp 55–60

  29. Hendrickx I, Kim SN, Kozareva Z, Nakov P, Séaghdha DO, Padó S, Pennacchiotti M, Romano L, Szpakowicz S (2010) Semeval-2010 task 8: multi-way classification of semantic relations between pairs of nominals. ACL 2010:33

    Google Scholar 

  30. Angeli G, Tibshirani J, Wu J, Manning CD (2014) Combining distant and partial supervision for relation extraction. In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), pp 1556–1567

  31. Guo X, Zhang H, Yang H, Xu L, Ye Z (2019) A single attention-based combination of CNN and RNN for relation classification. IEEE Access 7:12467–12475

    Article  Google Scholar 

  32. Yu B, Mengge X, Zhang Z, Liu T, Yubin W, Wang B (2020) Learning to prune dependency trees with rethinking for neural relation extraction. In: Proceedings of the 28th international conference on computational linguistics, pp 3842–3852

  33. Wang H, Qin K, Lu G, Luo G, Liu G (2020) Direction-sensitive relation extraction using Bi-SDP attention model. Knowl-Based Syst 198:105928

    Article  Google Scholar 

  34. Wang D, Tiwari P, Garg S, Zhu H, Bruza P (2020) Structural block driven enhanced convolutional neural representation for relation extraction. Appl Soft Comput 86:105913

    Article  Google Scholar 

  35. Li D, Lei Z-L, Song B-Y, Ji W-T, Kou Y (2022) Neural attentional relation extraction with dual dependency trees. J Comput Sci Technol 37(6):1369–1381

    Article  Google Scholar 

  36. Wang Y, Han Z, You K, Lin Z (2022) A two-channel model for relation extraction using multiple trained word embeddings. Knowl-Based Syst 255:109701

    Article  Google Scholar 

Download references

Acknowledgements

This work is supported in part by Postgraduate Education Reform and Quality Improvement Project of Henan Province under Grant YJS2022JD26; in part by Research and Practice Project of Postgraduate Education and Teaching Reform of Henan University under Grant YJSJG2022XJ039; and in part by Postgraduate Training Innovation and Quality Improvement Action Plan Project of Henan University under Grant (Talent Plan) SYLYC2022148 and SYLYC2023138, Grant (Education Innovation Training Base) SYLJD2022008, Grant (Case Library) SYLAL2023017. (Corresponding authors: Ying Wang)

Author information

Authors and Affiliations

Authors

Contributions

NL and YW made substantial contributions to the conception or design of the work. NL and YW conducted experiments and performed data analysis. NL, YW, and TL drafted the original manuscript. All authors reviewed the results and approved the final version of the manuscript.

Corresponding author

Correspondence to Ying Wang.

Ethics declarations

Conflict of interest

The authors declare no Conflict of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Li, N., Wang, Y. & Liu, T. Dependency-position relation graph convolutional network with hierarchical attention mechanism for relation extraction. J Supercomput 80, 18954–18976 (2024). https://doi.org/10.1007/s11227-024-06204-8

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11227-024-06204-8

Keywords