Skip to main content

Attention Adaptive Chinese Named Entity Recognition Based on Vocabulary Enhancement

  • Conference paper
Intelligent Information Processing XI (IIP 2022)

Part of the book series: IFIP Advances in Information and Communication Technology ((IFIPAICT,volume 643))

Included in the following conference series:

Abstract

To deal with the lack of word information in character vector embedding and the problem of Out-of-Vocabulary in Named Entity Recognition, an attention adaptive chinese named entity recognition (CNER) model based on vocabulary enhancement (ACVE) is proposed. The mechanism of potential information embedding is designed, which acquires word-level potential information by constructing semantic vectors, and the fusion embedding of character information and word-level information realizes the enhancement of semantic features; We also propose an attention mechanism for adaptive distribution, which adaptively adjusts the position of attention by introducing a dynamic scaling factor to obtain the attention distribution suitable for NER tasks. Experiments on a special field dataset with a large number of out-of-vocabulary (OOV) words show that, compared with state-of-the-art methods, our method is more effective and achieves better results.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 119.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 159.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 159.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. He, H., Sun, X.: A unified model for cross-domain and semi-supervised named entity recognition in chinese social media. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 31, no. 1 (2017)

    Google Scholar 

  2. Luo, W., Yang, F.: An empirical study of automatic chinese word segmentation for spoken language understanding and named entity recognition. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 238–248 (2016)

    Google Scholar 

  3. Zhang, Y., Yang, J.: Chinese NER using lattice LSTM. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), vol. 1, pp. 1554–1564 (2018)

    Google Scholar 

  4. Sui, D., Chen, Y., Liu, K., Zhao, J., Liu, S.: Leverage lexical knowledge for chinese named entity recognition via collaborative graph network. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing EMNLP-IJCNLP, pp. 3828–3838 (2019)

    Google Scholar 

  5. Gui, T., Zou, Y., Zhang, Q.: A lexicon-based graph neural network for Chinese NER. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 1040–1050 (2019)

    Google Scholar 

  6. Liu, W., Xu, T., Xu, Q., Song, J., Zu, Y.: An encoding strategy based word character LSTM for Chinese NER. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 2379–2389 (2019)

    Google Scholar 

  7. Devlin, J., et al.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), vol. 2019, pp. 4171–4186 (2019)

    Google Scholar 

  8. Yan, H., Deng, B., Li, X., Qiu, X.: TENER: adapting transformer encoder for named entity recognition. arXiv Preprint arXiv:1911.04474 (2019)

  9. Jia, C., Shi, Y., Yang, Q., Zhang, Y.: Entity enhanced BERT pre-training for Chinese NER. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), November 2020, pp. 6384–6396 (2020)

    Google Scholar 

  10. Lample, G., et al.: Neural architectures for named entity recognition. In: 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL HLT 2016-Proceedings of the Conference, pp. 260–270 (2016)

    Google Scholar 

  11. Huang, Z., Xu, W., Yu, K.: Bidirectional LSTM-CRF models for sequence tagging. arXiv preprint arXiv:1508.01991 (2015)

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Quansheng Dou .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 IFIP International Federation for Information Processing

About this paper

Cite this paper

Zhao, P., Dou, Q., Jiang, P. (2022). Attention Adaptive Chinese Named Entity Recognition Based on Vocabulary Enhancement. In: Shi, Z., Zucker, JD., An, B. (eds) Intelligent Information Processing XI. IIP 2022. IFIP Advances in Information and Communication Technology, vol 643. Springer, Cham. https://doi.org/10.1007/978-3-031-03948-5_32

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-03948-5_32

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-03947-8

  • Online ISBN: 978-3-031-03948-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics