Abstract
Multimodal tasks are gradually attracting the attention of the research community, and the lack of multimodal event extraction datasets restricts the development of multimodal event extraction. We introduce the new Multimodal Event Extraction Dataset (MEED) to fill the gap, we define event types and argument roles that can be used on multimodal data, then use controllable text generation to generate the textual modality based on visual event extraction dataset. In this paper, we aim to make full use of multimodal resources in the event extraction task by constructing a large-scale and high-quality multimodal event extraction dataset and promote researches in the field of multimodal event extraction.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Caba Heilbron, F., Escorcia, V., Ghanem, B., Carlos Niebles, J.: ActivityNet: a large-scale video benchmark for human activity understanding. In: CVPR, pp. 961–970 (2015)
Chen, S., Jin, Q., Wang, P., Wu, Q.: Say as you wish: fine-grained control of image caption generation with abstract scene graphs. In: CVPR, pp. 9962–9971 (2020)
Ebner, S., Xia, P., Culkin, R., Rawlins, K., Van Durme, B.: Multi-sentence argument linking. In: ACL, pp. 8057–8077 (2020)
Hu, Z., Yang, Z., Liang, X., Salakhutdinov, R., Xing, E.P.: Toward controlled generation of text. In: ICML, pp. 1587–1596 (2017)
Li, M., et al.: Cross-media structured common space for multimedia event extraction. In: ACL, pp. 2557–2568 (2020)
Nuij, W., Milea, V., Hogenboom, F., Frasincar, F., Kaymak, U.: An automated framework for incorporating news into stock trading strategies. IEEE Trans. Knowl. Data Eng. 26(4), 823–835 (2013)
Pratt, S., Yatskar, M., Weihs, L., Farhadi, A., Kembhavi, A.: Grounded situation recognition. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12349, pp. 314–332. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-58548-8_19
Vanegas, J., Matos, S., González, F., Oliveira, J.: An overview of biomolecular event extraction from scientific documents. Comput. Math. Methods Med. 2015, 571381 (2015)
Yang, S., Feng, D., Qiao, L., Kan, Z., Li, D.: Exploring pre-trained language models for event extraction and generation. In: ACL, pp. 5284–5294 (2019)
Yatskar, M., Zettlemoyer, L., Farhadi, A.: Situation recognition: visual semantic role labeling for image understanding. In: CVPR, pp. 5534–5542 (2016)
Acknowledgment
This work was supported by National Natural Science Foundation of China with Grant No. 61906037; the Fundamental Research Funds for the Central Universities with No. 224202k10011; the CCF-Baidu Open Fund with No. CCF BAIDU OF2020003.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Wang, S., Zheng, Q., Su, Z., Na, C., Qi, G. (2021). MEED: A Multimodal Event Extraction Dataset. In: Qin, B., Jin, Z., Wang, H., Pan, J., Liu, Y., An, B. (eds) Knowledge Graph and Semantic Computing: Knowledge Graph Empowers New Infrastructure Construction. CCKS 2021. Communications in Computer and Information Science, vol 1466. Springer, Singapore. https://doi.org/10.1007/978-981-16-6471-7_23
Download citation
DOI: https://doi.org/10.1007/978-981-16-6471-7_23
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-16-6470-0
Online ISBN: 978-981-16-6471-7
eBook Packages: Computer ScienceComputer Science (R0)