Abstract
Events are vital parts of natural language, reflecting the state changes of entities. The Event Extraction (EE) task aims to extract event triggers (the most representative words or phrases) and their arguments (participants in the event) from the given text. Most current works use sequence tagging models to solve the EE task. However, those methods only treat event and argument types as different class numbers, ignoring the semantics of those labels. However, label semantics are critical in the EE task. For example, the trigger word “fight” is semantically closer to the event type “Conflict:Attack” rather than “Life:Marriage”. To emphasize the label semantics in events, we formulate EE as a prototype matching task and propose a Prototype Matching framework for Joint Event Extraction (PMJEE). Specifically, prototypical embeddings for both event trigger and argument types are introduced to encode their label semantics and correlations. Then a dual-channel attention layer and extraction modules are used to jointly extract event triggers and arguments. Prototypical embeddings will be optimized during training to improve the event extraction performance. Extensive experiments indicate our method achieves better performance than strong baselines, especially in data-scarce scenarios. In the detailed analysis, we verify the effectiveness of each part of the model and explore the impact of different label semantic materials.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Chen, Y., Xu, L., Liu, K., Zeng, D., Zhao, J.: Event extraction via dynamic multi-pooling convolutional neural networks. In: Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 167–176 (2015)
Deng, S., Zhang, N., Kang, J., Zhang, Y., Zhang, W., Chen, H.: Meta-learning with dynamic-memory-based prototypical network for few-shot event detection. In: Proceedings of the 13th International Conference on Web Search and Data Mining, pp. 151–159 (2020)
Deng, S., et al.: OntoED: low-resource event detection with ontology embedding. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 2828–2839. Association for Computational Linguistics, Online (2021)
Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, Minnesota, pp. 4171–4186. Association for Computational Linguistics (2019)
Ding, N., et al.: Prototypical Representation Learning for Relation Extraction. arXiv:2103.11647 (2021)
Dong, L., et al.: Unified language model pre-training for natural language understanding and generation. In: Advances in Neural Information Processing Systems, vol. 32. Curran Associates, Inc. (2019)
Gao, T., Han, X., Liu, Z., Sun, M.: Hybrid attention-based prototypical networks for noisy few-shot relation classification. In: AAAI, vol. 33, pp. 6407–6414 (2019)
Lewis, M., et al.: Bart: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv preprint arXiv:1910.13461 (2019)
Li, H., Mo, T., Fan, H., Wang, J., Wang, J., Zhang, F., Li, W.: KiPT: knowledge-injected prompt tuning for event detection. In: Proceedings of the 29th International Conference on Computational Linguistics, pp. 1943–1952 (2022)
Lima, R., Espinasse, B., Freitas, F.: OntoILPER: an ontology- and inductive logic programming-based system to extract entities and relations from text. Knowl. Inf. Syst. 56(1), 223–255 (2018)
Lin, Y., Ji, H., Huang, F., Wu, L.: A joint neural model for information extraction with global features. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 7999–8009 (2020)
Liu, X., Luo, Z., Huang, H.: Jointly multiple events extraction via attention-based graph information aggregation. arXiv preprint arXiv:1809.09078 (2018)
Liu, Y., et al.: Roberta: a robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 (2019)
Lu, Y., et al.: Text2Event: controllable sequence-to-structure generation for end-to-end event extraction. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 2795–2806. Association for Computational Linguistics, Online (2021)
Nguyen, T.H., Cho, K., Grishman, R.: Joint event extraction via recurrent neural networks. In: Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, San Diego, California, pp. 300–309. Association for Computational Linguistics (2016)
Nguyen, T.M., Nguyen, T.H.: One for all: neural joint modeling of entities and events. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, no. 01, pp. 6851–6858 (2019)
Paolini, G., et al.: Structured prediction as translation between augmented natural languages. arXiv preprint arXiv:2101.05779 (2021)
Snell, J., Swersky, K., Zemel, R.: Prototypical networks for few-shot learning. In: Advances in Neural Information Processing Systems, vol. 30 (2017)
Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, vol. 30 (2017)
Wadden, D., Wennberg, U., Luan, Y., Hajishirzi, H.: Entity, relation, and event extraction with contextualized span representations. arXiv preprint arXiv:1909.03546 (2019)
Yan, H., Jin, X., Meng, X., Guo, J., Cheng, X.: Event detection with multi-order graph convolution and aggregated attention. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, pp. 5766–5770. Association for Computational Linguistics (2019)
Yang, S., Feng, D., Qiao, L., Kan, Z., Li, D.: Exploring pre-trained language models for event extraction and generation. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, pp. 5284–5294. Association for Computational Linguistics (2019)
Acknowledgment
This work was supported in part by National Key R &D Program of China under Grants No. 2022YFF0902703.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Li, H., Mo, T., Geng, D., Li, W. (2023). PMJEE: A Prototype Matching Framework for Joint Event Extraction. In: Wang, X., et al. Database Systems for Advanced Applications. DASFAA 2023. Lecture Notes in Computer Science, vol 13946. Springer, Cham. https://doi.org/10.1007/978-3-031-30678-5_2
Download citation
DOI: https://doi.org/10.1007/978-3-031-30678-5_2
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-30677-8
Online ISBN: 978-3-031-30678-5
eBook Packages: Computer ScienceComputer Science (R0)