Abstract
Event extraction is an important task in the field of natural language processing. However, most of the existing event extraction techniques focus on sentence-level extraction, which inevitably ignores the contextual features of sentences and the occurrence of multiple event trigger words in the same sentence. Therefore, this paper mainly uses the multi-head self-attention mechanism to integrate text features from multiple dimensions and levels to achieve the task of event detection at the level of text. First, convolutional neural network combined with dynamic multi-pool strategy is used to extract sentence level features. Secondly, the discourse feature representation of full-text information fusion is obtained by multi-head self-attention mechanism model. Finally, using the classifier function to classify, and then detect the trigger word and category of the event. Experimental results show that the proposed method achieves good results in document-level event extraction.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Hoogenboom, F.P., Flavius, F., Uzay, K., Franciska, D.J., Caron, E.A.M.: A survey of event extraction methods from text for decision support systems. Decis. Support Syst. 85(1), 12–22 (2016)
Ji, S., Pan, S., Cambria, E., Marttinen, P., Yu, P.S.: A survey on knowledge graphs: representation, acquisition, and applications. Trans. Neural Netw. Learn. Syst. 33(2), 494–514 (2021)
Garg, S., Vu, T., Moschitti, A.: TANDA: transfer and adapt pre-trained transformer models for answer sentence selection. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, no. 5, pp. 7780–7788 (2020)
Xiang, W., Wang, B.: A survey of event extraction from text. Access 7, 173111–173137 (2019)
Gao, L.Z., Gang, Z., Luo, J.Y., Lan, M.J.: Survey on meta-event extraction. Comput. Sci. 46(8), 9–15 (2019)
Wen, H.Y., Qu, Y.R., Ji, H., Han, J.W., et al.: Event time extraction and propagation via graph attention networks. In: Human Language Technologies, pp. 62–73 (2021)
Du, X.Y., Rush, A., Cardie, C.: GRIT: generative role-filler transformers for document-level event entity extraction (2020)
Walker, C., Strassel, S., Medero, J., Maeda, K.: ACE 2005 multilingual training corpus. Linguistic Data Consortium, Philadelphia (2006)
Chen, Y.B., Xu, L.H., Liu, K., Zeng, D.J., Zhao, J.: Event extraction via dynamic multi-pooling convolutional neural networks. In: Association for Computational Linguistics, pp. 167–176 (2015)
Vaswani, A., Shazeer, N., Parmar, M., Uszkoreit, J., Jones, L., et al.: Attention is all you need. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, no. 11, pp. 6000–6010 (2017)
Davani, A.M., Yeh, L., et al.: Reporting the unreported: event extraction for analyzing the local representation of hate crimes. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, pp. 5753–5757 (2019)
Liu, X., Huang, H.Y., Zhang, Y.: Open domain event extraction using neural latent variable models. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 2860–2871 (2019)
Yang, Y.M., Pierce, T., Jaime, G.C.: A study of retrospective and online event detection. In: SIGIR, p. 98 (1998)
Kim, J.T., Moldovan, D.I.: PALKA: a system for lexical knowledge acquisition. In: Proceedings of the Second International Conference on Information and Knowledge Management, no. 8, pp. 124–131 (1993)
Suthaharan, S.: Support vector machine. In: Machine Learning Models and Algorithms for Big Data Classification Integrated Series in Information Systems, vol. 36 (2006)
Daniel, L., Pedro, D.: Naive Bayes models for probability estimation. In: Proceedings of the 22nd International Conference on Machine Learning, no. 8, pp. 529–536 (2005)
Eddy, S.R.: What is a hidden Markov model? Nat. Biotechnol. 22(10), 1315–1316 (2004)
Berger, A., Della Pietra, S.A., Della Pietra, V.J.: A maximum entropy approach to natural language processing. Comput. Linguist. 22(33), 39–71 (1996)
Nguyen, T.M., Nguyen, T.H.: One for all: neural joint modeling of entities and events. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, no. 1, pp. 6851–6858 (2019)
Feng, X., Qin, B., Liu, T.: A language-independent neural network for event detection. SCIENCE CHINA Inf. Sci. 61(9), 1–12 (2018). https://doi.org/10.1007/s11432-017-9359-x
Yang, H., Chen, Y.B., Liu, K., Xiao, Y., Zhao, J.: DCFEE: a document-level Chinese financial event extraction system based on automatically labeled training data. In: Proceedings of ACL 2018 System Demonstrations, pp. 50–55 (2018)
Wang, X., Wang, Z., Han, X., et al.: MAVEN: a massive general domain event detection dataset. arXiv preprint. arXiv:2004.13590 (2020)
Acknowledgement
The authors would like to thank the associate editor and the reviewers for their time and effort provided to review the manuscript.
Funding
This work is supported by State Grid Shandong Electric Power Company Science and Technology Project Funding under Grant no. 62061320C007, SGSDWH00YXJS2000128, the Fundamental Research Funds for the Central Universities (Grant No. HIT. NSRIF.201714), Weihai Science and Technology Development Program (2016DX GJMS15), Weihai Scientific Research and Innovation Fund (2020) and Key Research and Development Program in Shandong Provincial (2017GGX90103).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Ethics declarations
The authors declare that they have no conflicts of interest to report regarding the present study.
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Qiao, X. et al. (2023). A Event Extraction Method of Document-Level Based on the Self-attention Mechanism. In: Xu, Y., Yan, H., Teng, H., Cai, J., Li, J. (eds) Machine Learning for Cyber Security. ML4CS 2022. Lecture Notes in Computer Science, vol 13656. Springer, Cham. https://doi.org/10.1007/978-3-031-20099-1_50
Download citation
DOI: https://doi.org/10.1007/978-3-031-20099-1_50
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-20098-4
Online ISBN: 978-3-031-20099-1
eBook Packages: Computer ScienceComputer Science (R0)