skip to main content
10.1145/3582197.3582205acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicitConference Proceedingsconference-collections
research-article

End-to-End Entity Linking Combined with Bert-based Siamese and Interaction Network

Published: 30 March 2023 Publication History

Abstract

The Entity Linking (EL) task is actually more complex than it seems. Not only are most of the given texts short and unlabeled with candidate mention information, but the entities to be predicted generally do not appear in the training set of the pre-trained model. This paper introduces an end-to-end Entity Linking model combined with Bert-based Siamese and Interaction Network, which can complete the EL task quickly and accurately given only short textual information without pre-specified mention boundaries at the time of input. The model is divided into two stages to achieve end-to-end entity linking. The first part is a pretrained BERT encoder-based Siamese network (BSN) architecture. BSN includes a text encoder and an entity encoder. These encoders are used to obtain the word embedding of the input text and the entity description token of the knowledge base respectively. We can also calculate the probability that a span can be used as a mention. In the candidate generation process, we first take the average value of the token word vector in a mention as the representation vector of the mention. We further obtain the similarity between the mention and the entity in the knowledge base and the span corresponding to an entity probability. During training, the loss of mention detection and candidate generation is counted as the loss of the first part. In the second part, we add an interaction network (BIN) based on pretrained BERT encoder (BIN). Concatenate the mention context and candidate entities, obtain the representation vector of the mention context-candidate entity through the encoder. Then add a linear layer to obtain the score and reorder the original candidate entities to achieve final disambiguation. Besides, a corresponding solution is proposed for the Speed-Accuracy Tradeoff problem. In experiments, the model achieves near-SOTA performance on the Zero-shot EL dataset and the WikilinksNED dataset. And it is shown that the stage of Mention Detection (MD) and Candidate Entity Generation (CEG) are crucial in the end-to-end entity linking process, and the two tasks can benefit from each other from joint training.

References

[1]
Sevgili, Özge, "Neural entity linking: A survey of models based on deep learning." Semantic Web Preprint (2022): 1-44.
[2]
Yamada, Ikuya, "Global entity disambiguation with BERT." Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 2022.
[3]
Yang, Yi, and Ming-Wei Chang. "S-MART: Novel Tree-based Structured Learning Algorithms Applied to Tweet Entity Linking." Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). 2015.
[4]
Zhang, Wenzheng, Wenyue Hua, and Karl Stratos. "EntQA: Entity Linking as Question Answering." International Conference on Learning Representations. 2021.
[5]
Logeswaran, Lajanugen, "Zero-Shot Entity Linking by Reading Entity Descriptions." Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 2019.
[6]
Nguyen, Dat Ba, Martin Theobald, and Gerhard Weikum. "J-NERD: Joint Named Entity Recognition and Disambiguation with Rich Linguistic Features." Transactions of the Association for Computational Linguistics 4. 2016. 215-229.
[7]
Raiman, Jonathan Raphael, and Olivier Michel Raiman. "DeepType: Multilingual Entity Linking by Neural Type System Evolution." Thirty-Second AAAI Conference on Artificial Intelligence. 2018.
[8]
Le, Phong, and Ivan Titov. "Improving Entity Linking by Modeling Latent Relations between Mentions." Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2018.
[9]
Wu, Ledell, "Scalable Zero-shot Entity Linking with Dense Entity Retrieval." Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP). 2020.
[10]
Kolitsas, Nikolaos, Octavian-Eugen Ganea, and Thomas Hofmann. "End-to-End Neural Entity Linking." Proceedings of the 22nd Conference on Computational Natural Language Learning. 2018.
[11]
Lee, Daniel D., Advances in neural information processing systems 22. Tech. Rep., Tech. Rep, 2009.
[12]
Kenton, Jacob Devlin Ming-Wei Chang, and Lee Kristina Toutanova. "Bert: Pre-training of deep bidirectional transformers for language understanding." Proceedings of naacL-HLT. 2019.
[13]
Humeau, Samuel, "Poly-encoders: Architectures and Pre-training Strategies for Fast and Accurate Multi-sentence Scoring." International Conference on Learning Representations. 2019.
[14]
Eshel, Yotam, "Named entity disambiguation for noisy text." arXiv preprint arXiv:1706.09147. 2017.
[15]
Guo, Zhaochen, and Denilson Barbosa. "Robust named entity disambiguation with random walks." Semantic Web 9.4. 2018.
[16]
Broscheit, Samuel. "Investigating entity knowledge in BERT with simple neural end-to-end entity linking." arXiv preprint arXiv:2003.05473. 2020.
[17]
Li, Belinda Z., "Efficient One-Pass End-to-End Entity Linking for Questions." Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP). 2020.
[18]
Gillick, Dan, "Learning Dense Representations for Entity Retrieval." Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL). 2019.

Index Terms

  1. End-to-End Entity Linking Combined with Bert-based Siamese and Interaction Network
          Index terms have been assigned to the content through auto-classification.

          Recommendations

          Comments

          Information & Contributors

          Information

          Published In

          cover image ACM Other conferences
          ICIT '22: Proceedings of the 2022 10th International Conference on Information Technology: IoT and Smart City
          December 2022
          385 pages
          ISBN:9781450397438
          DOI:10.1145/3582197
          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          Published: 30 March 2023

          Permissions

          Request permissions for this article.

          Check for updates

          Author Tags

          1. Bert-based Interaction Network
          2. Bert-based Siamese Network
          3. Entity Disambiguation
          4. Entity Linking
          5. end-to-end

          Qualifiers

          • Research-article
          • Research
          • Refereed limited

          Conference

          ICIT 2022
          ICIT 2022: IoT and Smart City
          December 23 - 25, 2022
          Shanghai, China

          Contributors

          Other Metrics

          Bibliometrics & Citations

          Bibliometrics

          Article Metrics

          • 0
            Total Citations
          • 49
            Total Downloads
          • Downloads (Last 12 months)11
          • Downloads (Last 6 weeks)1
          Reflects downloads up to 01 Mar 2025

          Other Metrics

          Citations

          View Options

          Login options

          View options

          PDF

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader

          HTML Format

          View this article in HTML Format.

          HTML Format

          Figures

          Tables

          Media

          Share

          Share

          Share this Publication link

          Share on social media