Skip to main content
Log in

A fine-grained causality extraction model incorporating relative location coding

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

Popular methods of causality extraction work well for simple and explicit single causal relations, but it remains challenging to extract causal relations from the complex sentences of natural texts due to ambiguity concerning the locations of the causal subject and object as well as the complexity of the relevant dependencies. To solve these problems, this paper proposes a five-tuple annotation scheme that defines a scoring function to iteratively parse out the causal pairs from multiple entity pairs in a sentence, and to thus transform the task of extracting causal relations into that of automatic annotation. First, this study uses this scheme to propose a multi-headed, self-attentive mechanism that incorporates encoded information on relative position to increase the capability of the model to perceive causal features. Second, the authors combine information from a dependency tree while assigning the appropriate weights, and finally use a bidirectional GCN network to parse the weights of features of the tree from multiple perspectives and splice the dependency-related features. This joint model of extraction improves the bounds of cause–effect pairs of entities while considering the dependency relationships between them, which renders the extracted, fine-grained causal terms more accurate. Experiments on the SemEval 2010 task 8 and the ADE datasets show that our approach significantly outperforms prevalent methods in terms of the accuracy of solving complex causal extraction compared with state-of-the-art approaches to modeling.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15

Similar content being viewed by others

Data availability

The data sets supporting the results of this article are included within the article and its additional files.

References

  1. Radinsky K, Davidovich S, Markovitch S (2012) Learning causality for news events prediction. In: Proceedings of the 21st International Conference on World Wide Web, pp 909–918

  2. Pechsiri C, Kawtrakul A (2007) Mining causality from texts for question answering system. IEICE Trans Inf Syst 90(10):1523–1533

    Article  Google Scholar 

  3. Jun EJ, Bautista AR, Nunez MD, Allen DC, Tak JH, Alvarez E, Basso MA (2021) Causal role for the primate superior colliculus in the computation of evidence for perceptual decisions. Nat Neurosci 24(8):1121–1131

    Article  Google Scholar 

  4. Lee D-G, Shin H (2017) Disease causality extraction based on lexical semantics and document-clause frequency from biomedical literature. BMC Med Inform Decis Mak 17(1):1–9

    MathSciNet  Google Scholar 

  5. Xu Y, Liu J (2021) High-speed train fault detection with unsupervised causality-based feature extraction methods. Adv Eng Inform 49:101312

    Article  Google Scholar 

  6. Garcia, D (1997) Coatis, an nlp system to locate expressions of actions connected by causality links. In: International conference on knowledge engineering and knowledge management. Springer, pp 347–352

  7. Zhao S, Liu T, Zhao S, Chen Y, Nie J-Y (2016) Event causality extraction based on connectives analysis. Neurocomputing 173:1943–1950

    Article  Google Scholar 

  8. Kim HD, Castellanos M, Hsu M, Zhai C, Rietz T, Diermeier D (2013) Mining causal topics in text data: iterative topic modeling with time series feedback. In: Proceedings of the 22nd ACM International Conference on Information Knowledge Management, pp 885–890

  9. Lin Z, Kan M-Y, Ng HT (2009) Recognizing implicit discourse relations in the penn discourse treebank. In: Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing, pp 343–351

  10. Li F, Zhang M, Fu G, Ji D (2017) A neural joint model for entity and relation extraction from biomedical text. BMC Bioinformatics 18(1):1–11

    Article  Google Scholar 

  11. Wang J, Lu W (2020) Two are better than one: Joint entity and relation extraction with table-sequence encoders. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 1706–1721

  12. Li Z, Li Q, Zou X, Ren J (2021) Causality extraction based on self-attentive bilstm-crf with transferred embeddings. Neurocomputing 423:207–219

    Article  Google Scholar 

  13. Wang L, Cao Z, De Melo G, Liu Z (2016) Relation classification via multilevel attention cnns. In: Proceedings of the 54th annual meeting of the association for computational linguistics (Volume 1: Long Papers), pp 1298–1307

  14. Wang G, Liu S, Wei F (2022) Weighted graph convolution over dependency trees for nontaxonomic relation extraction on public opinion information. Appl Intell 52(3):3403–3417

    Article  Google Scholar 

  15. Xu Y, Mou L, Li G, Chen Y, Peng H, Jin Z (2015) Classifying relations via long short term memory networks along shortest dependency paths. In: Proceedings of the 2015 conference on empirical methods in natural language processing, pp 1785–1794

  16. Tuo M, Yang W, Wei F, Dai Q (2023) A novel chinese overlapping entity relation extraction model using word-label based on cascade binary tagging. Electronics 12(4):1013

    Article  Google Scholar 

  17. Zhang Y, Zhong V, Chen D, Angeli G, Manning CD (2017) Position-aware attention and supervised data improve slot filling. In: Conference on empirical methods in natural language processing

  18. Dasgupta T, Saha R, Dey L, Naskar A (2018) Automatic extraction of causal relations from text using linguistically informed deep neural networks. In: Proceedings of the 19th annual SIGdial meeting on discourse and dialogue, pp 306–316

  19. Yuan C, Fan C, Bao J, Xu R (2020) Emotion-cause pair extraction as sequence labeling based on a novel tagging scheme, 3568–3573

  20. De Marneffe M-C, Manning CD (2008) Stanford typed dependencies manual. Report, Technical report, Stanford University

  21. Lee J, Lee I, Kang J (2019) Self-attention graph pooling. In: International conference on machine learning. PMLR, pp 3734–3743

  22. Shaw P, Uszkoreit J, Vaswani A (2018) Self-attention with relative position representations. In: Proceedings of NAACL-HLT, pp. 464–468

  23. Fu S, Liu W, Zhang K, Zhou Y, Tao D (2021) Semi-supervised classification by graph p-laplacian convolutional networks. Inf Sci 560:92–106

    Article  MathSciNet  Google Scholar 

  24. Fu T-J, Li P-H, Ma W-Y (2019) Graphrel: Modeling text as relational graphs for joint entity and relation extraction. In: Proceedings of the 57th annual meeting of the association for computational linguistics, pp 1409–1418

  25. Jin S, Jang H, Kim W (2018) Improving bidirectional lstm-crf model of sequence tagging by using ontology knowledge based feature. J Intell Inf Syst 24(1):253–266

    Google Scholar 

  26. Bekoulis G, Deleu J, Demeester T, Develder C (2018) Joint entity recognition and relation extraction as a multi-head selection problem. Expert Syst Appl 114:34–45

    Article  Google Scholar 

Download references

Acknowledgements

The authors gratefully acknowledge the financial supports by the National Key R&D Program of China (Grant No. 2020AAA0109300).

Funding

The research leading to these results received funding from the National Key R&D Program of China under Grant Agreement Grant No. 2020AAA0109300.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Weibing Wan.

Ethics declarations

Conflict of interest

The authors have no conflicts of interest to declare that are relevant to the content of this article.

Additional information

Publisher's note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Wan, W., Chen, Y., Gao, Y. et al. A fine-grained causality extraction model incorporating relative location coding. Appl Intell 53, 27163–27176 (2023). https://doi.org/10.1007/s10489-023-04970-1

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10489-023-04970-1

Keywords

Navigation