Abstract
As an important type of relationship, causality plays a vital role in relation reasoning. Therefore, causality extraction from natural language texts is a crucial task, especially in the field of traffic. For example, we can quickly discover the causes of traffic accidents and the correlation between the cause events. Existing methods utilize machine learning models to extract simple causality, however, there is nested causality in the traffic accident sentences, which is important for us to reason the cause and effect. In order to extract the causality successfully, we simplify the complex nested causality structure to the pairwise causality structure. On this basis, we propose a method that contains two steps. First, we extract the cause events from the input sentence, second, combine the extracted cause with the incomplete question template to obtain a complete question sentence, then we adopt the way of question answering tasks to extract the effect events. Experiments on the traffic accident dataset show the effectiveness of our model. However, we observe that due to the small training set, there is still room for improvement in the extraction accuracy of nested causality.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Chen, D., Cao, Y., Luo, P.: Pairwise causality structure: towards nested causality mining on financial statements. In: Zhu, X., Zhang, M., Hong, Yu., He, R. (eds.) NLPCC 2020. LNCS (LNAI), vol. 12430, pp. 725–737. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-60450-9_57
Cao, Y., Chen, D., Li, H., Luo, P.: Nested relation extraction with iterative neural network. In: CIKM, pp. 1001–1010. ACM (2019)
Xu, J., Zuo, W., Liang, S., Wang, Y.: Causal relation extraction based on graph attention networks. J. Comput. Res. Dev. 57(1), 159–174 (2020)
Zheng, Q., Wu, Z., Zuo, J.: Event causality extraction based on two-layer CNN-BiGRU-CRF model. Comput. Eng. 47(05), 58–64 (2021)
Li, Z., Li, Q., Zou, X., Ren, J.: Causality extraction based on self-attentive BiLSTM-CRF with transferred embeddings. Neurocomputing 423, 207–219 (2021)
Wei, Z., Su, J., Wang, Y., Tian, Y., Yi, C.: A novel cascade binary tagging framework for relational triple extraction. In: ACL, pp. 1476–1488 (2020)
Li, X., Yin, F., Sun, Z., et al.: Entity-relation extraction as multi-turn question answering. In: ACL, pp. 1340–1350 (2019)
Devlin, J., Chang, M., Lee, K., Toutanova, K.: Bert: pre-training of deep bidirectional transformers for language understanding. In: NAACL-HLT, pp. 4171–4186 (2019)
Zhong, P., Wang, D, Miao, C.: Knowledge-enriched transformer for emotion detection in textual conversations. In: EMNLP-IJCNLP, pp. 165–176 (2019)
Wei, J., Zou, K.: EDA: easy data augmentation techniques for boosting performance on text classification tasks. In: EMNLP-IJCNLP, pp. 6382–6388 (2019)
Li, X., Feng, J., Meng, Y., et al.: A unified MRC framework for named entity recognition. In: ACL, pp. 5849–5859 (2020)
Gu, Y., Yang, M., Lin, P.: Lightweight multiple perspective fusion with information enrich for BERT-based answer selection. In: NLPCC 2020, pp. 543–554(2020)
Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)
Acknowledgement
This work was supported by the NSFC (No. 61803337).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Zhou, G., Ma, W., Gong, Y., Wang, L., Li, Y., Zhang, Y. (2021). Nested Causality Extraction on Traffic Accident Texts as Question Answering. In: Wang, L., Feng, Y., Hong, Y., He, R. (eds) Natural Language Processing and Chinese Computing. NLPCC 2021. Lecture Notes in Computer Science(), vol 13029. Springer, Cham. https://doi.org/10.1007/978-3-030-88483-3_28
Download citation
DOI: https://doi.org/10.1007/978-3-030-88483-3_28
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-88482-6
Online ISBN: 978-3-030-88483-3
eBook Packages: Computer ScienceComputer Science (R0)