skip to main content
10.1145/3672919.3672991acmotherconferencesArticle/Chapter ViewAbstractPublication PagescsaideConference Proceedingsconference-collections
research-article

An InterFormer-BERT Chinese named entity recognition model with enhanced boundary

Published: 24 July 2024 Publication History

Abstract

Chinese Named Entity Recognition (NER) is more challenging compared to English, mainly because Chinese lacks clear word boundaries and natural separators. Certainly! In recent times, there has been a trend in utilizing pre-trained models augmented with adapter modules to create novel models for diverse NLP tasks, showcasing improved effectiveness. In this paper, we propose a boundary enhancement approach to improve Chinese NER. on the one hand, the representation of the internal dependencies of phrases is enhanced by an additional InterFormer (Inter Attention Network), which is then integrated as an adapter directly between the Transformer layers of BERT, which helps to enable a closer integration of word boundaries and semantic information, and on the other hand, to adding special tokens at the head and tail of entities (i.e., boundaries) as a supplementary task, this paper presents a novel framework for learning boundary information and performing entity recognition jointly. Experimental results on the Weibo, Resume, OntoNotes, and MRSA corpora validate the effectiveness of the proposed approach.

References

[1]
Luo Y, Xiao F, Zhao H. Hierarchical contextualized representation for named entity recognition[C]//Proceedings of the AAAI conference on artificial intelligence. 2020, 34 (05): 8441-8448.
[2]
Li X, Yan H, Qiu X, FLAT: Chinese NER Using Flat-Lattice Transformer[J]. 2020.
[3]
Wu S, Song X, Feng Z, NFLAT: Non-Flat-Lattice Transformer for Chinese Named Entity Recognition[J]. 2022.
[4]
Liu W, Fu X, Zhang Y, Lexicon Enhanced Chinese Sequence Labeling Using BERT Adapter[J]. 2021.
[5]
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language under-standing. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186, Minneapolis, Minnesota. Association for Computational Linguistics
[6]
Neil Houlsby, Andrei Giurgiu, Stanislaw Jastrzebski, Bruna Morrone, Quentin De Laroussilhe, Andrea Gesmundo, Mona Attariyan, and Sylvain Gelly. 2019. Parameter-efficient transfer learning for NLP. In Proceedings of the 36th International Conference on Machine Learning.
[7]
Nanyun Peng and Mark Dredze. 2015. Named entity recognition for Chinese social media with jointly trained embeddings. In Proceedings of the 2015 Conference on Empirical Methods in Natural Lan-guage Processing, pages 548–554, Lisbon, Portugal. Association for Computational Linguistics
[8]
Yue Zhang and Jie Y ang. 2018. Chinese NER using lattice LSTM. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (V olume 1: Long Papers), pages 1554–1564, Melbourne, Australia. Association for Computational Linguistics.
[9]
Gina-Anne Levow. 2006. The third international Chinese language processing bakeoff: Word segmentation and named entity recognition. In Proceedings of the Fifth SIGHAN Workshop on Chinese Language Processing, pages 108–117, Sydney, Australia. Association for Computational Linguistics.
[10]
Ralph M Weischedel and Linguistic Data Consortium. 2013. Ontonotes release 5.0. Title from disc label.
[11]
R. Ma, M. Peng, Q. Zhang, Z. Wei, and X. Huang, “Simplify the usage of lexicon in Chinese NER,” in Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Online: Association for Computational Linguistics, Jul. 2020, pp. 5951–5960. [Online]. Available: https://aclanthology.org/2020.acl-main.528
[12]
Yan Song, Shuming Shi, Jing Li, and Haisong Zhang.2018. Directional skip-gram: Explicitly distinguishing left and right context for word embeddings. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers), pages 175–180, New Orleans, Louisiana. Association for Computational Linguistics.

Index Terms

  1. An InterFormer-BERT Chinese named entity recognition model with enhanced boundary

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    CSAIDE '24: Proceedings of the 2024 3rd International Conference on Cyber Security, Artificial Intelligence and Digital Economy
    March 2024
    676 pages
    ISBN:9798400718212
    DOI:10.1145/3672919
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 24 July 2024

    Permissions

    Request permissions for this article.

    Check for updates

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Conference

    CSAIDE 2024

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • 0
      Total Citations
    • 20
      Total Downloads
    • Downloads (Last 12 months)20
    • Downloads (Last 6 weeks)3
    Reflects downloads up to 05 Mar 2025

    Other Metrics

    Citations

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format.

    HTML Format

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media