Abstract
Recently, using lexicon information to improve the performance of Chinese named entity recognition has been proven to be effective. Moreover, the lexicon-based method represented by Lattice-LSTM has also become the mainstream. Although Lattice-LSTM can introduce lexicon information into characters to augment named entity recognition performance, it cannot make good use of unlabeled data, which contains abundant semantic information to assist the network to improve effect. And because Lattice-LSTM introduces much lexicon information, there is currently no suitable way to assign weights to each word. In this work, we propose a method that can effectively introduce lexicon information, which is also simple to implement and can be applied to various networks. Based on the lexicon method, this method uses external unlabeled data to count the word frequency and improved mutual information to represent the weight of the word to introduce lexicon information. And attention mechanism is used to dynamically assign weights to each part of lexicon information. In this method, the fusion of character and lexicon information is processed before the input layer, so that the method has a faster training speed and better versatility. Compared with other methods that are based on lexicon information, this method introduces additional prior knowledge, namely unlabeled data, and achieves better results when the scale of dataset is small. And when combined with the pre-trained language model, the performance is better (the F1 scores on Weibo dataset and Resume dataset are 96.73% and 71.53% respectively). Experimental research shows that our method surpasses many other excellent baseline methods in training speed and performance on two small-scale public Chinese named entity recognition datasets.
Similar content being viewed by others
Data availability
The data in this paper can be obtained by contacting the corresponding author’s email address.
Code availability
The code in this paper can be obtained by contacting the corresponding author’s email address.
References
Ali A, Zhu Y, Chen Q, Yu J, Cai H (2019) Leveraging spatio-temporal patterns for predicting citywide traffic crowd flows using deep hybrid neural networks. In: 2019 IEEE 25th International Conference on Parallel and Distributed Systems (ICPADS). IEEE, pp 125–132
Ali A, Zhu Y, Zakarya M (2021) A data aggregation based approach to exploit dynamic spatio-temporal correlations for citywide crowd flows prediction in fog computing. Multimed Tools Appl 80(20):31401–31433
Ali A, Zhu Y, Zakarya M (2021) Exploiting dynamic spatio-temporal correlations for citywide traffic flow prediction using attention based neural networks. Inf Sci 577:852–870
Ali A, Zhu Y, Zakarya M (2022) Exploiting dynamic spatio-temporal graph convolutional neural networks for citywide traffic flows prediction. Neural Netw 145:233–247
Arora R, Tsai CT, Tsereteli K, Kambadur P, Yang Y (2019) A semi-Markov structured support vector machine model for high-precision named entity recognition. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp 5862–5866
Cao P, Chen Y, Liu K, Zhao J (2020) Chinese named entity recognition via adaptive multi-pass memory network with hierarchical tagging mechanism. In: China National Conference on Chinese Computational Linguistics. Springer, Cham, pp 144–158
Chen Y, Xu L, Liu K, Zeng D, Zhao J (2015) Event extraction via dynamic multi-pooling convolutional neural networks. In: Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing, vol 1: Long Papers, pp 167–176
Chen Y, Wu C, Qi T, Yuan Z, Huang Y (2020) Named entity recognition in multi-level contexts. In: Proceedings of the 1st Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 10th International Joint Conference on Natural Language Processing, pp 181–190
Collobert R, Weston J, Bottou L, Karlen M, Kavukcuoglu K, Kuksa P (2011) Natural language processing (almost) from scratch. J Mach Learn Res 12(ARTICLE):2493–2537
Devlin J, Chang MW, Lee K, Toutanova K (2018) Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805
Dong C, Zhang J, Zong C, Hattori M, Di H (2016)Character-based LSTM-CRF with radical-level features for Chinese named entity recognition. Natural Language Understanding and Intelligent Applications. Springer, Cham, pp 239–250
Gui T, Zou Y, Zhang Q, Peng M, Fu J, Wei Z, Huang XJ (2019) A lexicon-based graph neural network for chinese ner. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp 1039–1049
Gui T, Ma R, Zhang Q, Zhao L, Jiang YG, Huang X (2019)CNN-Based Chinese NER with Lexicon Rethinking. In IJCAI, pp 4982–4988
Hammerton J (2003) Named entity recognition with long short-term memory. In Proceedings of the seventh conference on Natural language learning at HLT-NAACL 2003, pp 172–175
Huang Z, Xu W, Yu K (2015) Bidirectional LSTM-CRF models for sequence tagging. arXiv preprintarXiv:1508.01991
Jia Y, Ma X (2019), April Attention in Character-Based BiLSTM-CRF for Chinese Named Entity Recognition. In: Proceedings of the 2019 4th International Conference on Mathematics and Artificial Intelligence, pp 1–4
Jie Z, Xie P, Lu W, Ding R, Li L (2019) Better modeling of incomplete annotations for named entity recognition. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, vol 1 (Long and Short Papers), pp 729–734
Lample G, Ballesteros M, Subramanian S, Kawakami K, Dyer C (2016) Neural architectures for named entity recognition. arXiv preprint arXiv:1603.01360
Li X, Yan H, Qiu X, Huang XJ (2020) FLAT: Chinese NER Using Flat-Lattice Transformer. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp 6836–6842
Liu A, Du J, Stoyanov V (2019)Knowledge-augmented language model and its application to unsupervised named-entity recognition. arXiv preprint arXiv:1904.04458
Ma R, Peng M, Zhang Q, Wei Z, Huang XJ (2020) Simplify the usage of lexicon in Chinese NER. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp 5951–5960
Nie Y, Tian Y, Wan X, Song Y, Dai B (2020) Named entity recognition for social media texts with semantic augmentation. arXiv preprint arXiv:2010.15458
Peng N, Dredze M (2016) Improving named entity recognition for chinese social media with word segmentation representation learning. arXiv preprint arXiv:1603.00786
Riedel S, Yao L, McCallum A, Marlin BM (2013) Relation extraction with matrix factorization and universal schemas. In: Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp 74–84
Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Polosukhin I (2017) Attention is all you need. arXiv preprint arXiv:1706.03762
Wang Z, Shang J, Liu L, Lu L, Liu J, Han J (2019) Crossweigh: Training named entity tagger from imperfect annotations. arXiv preprint arXiv:1909.01441
Wu Y, Schuster M, Chen Z, Le QV, Norouzi M, Macherey W, Dean J (2016) Google’s neural machine translation system: Bridging the gap between human and machine translation. arXiv preprint arXiv:1609.08144
Yan H, Deng B, Li X, Qiu X (2019) Tener: Adapting transformer encoder for named entity recognition. arXiv preprint arXiv:1911.04474
Yang Y, Katiyar A (2020) Simple and effective few-shot named entity recognition with structured nearest neighbor learning. arXiv preprint arXiv:2010.02405
Zhang Y, Yang J (2018) Chinese NER using lattice LSTM. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, vol 1: Long Papers, pp 1554–1564
Zheng C, Cai Y, Xu J, Leung HF, Xu G (2019) A boundary-aware neural model for nested named entity recognition. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp 357–366
Zhou JT, Zhang H, Jin D, Zhu H, Fang M, Goh RSM, Kwok K (2019) Dual adversarial neural transfer for low-resource named entity recognition. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp 3461–3471
Author information
Authors and Affiliations
Contributions
Conceptualization: [Shaobin Huang], [Yongpeng Sha], [Rongsheng Li]; Methodology: [Yongpeng Sha], [Rongsheng Li]; Formal analysis and investigation: [Yongpeng Sha], [Rongsheng Li]; Writing - original draft preparation: [Yongpeng Sha]; Writing - review and editing: [Yongpeng Sha], [Rongsheng Li]; Supervision: [Shaobin Huang].
Corresponding author
Ethics declarations
Conflict of interest
The authors have no relevant financial or non-financial interests to disclose.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Huang, S., Sha, Y. & Li, R. A chinese named entity recognition method for small-scale dataset based on lexicon and unlabeled data. Multimed Tools Appl 82, 2185–2206 (2023). https://doi.org/10.1007/s11042-022-13377-y
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-022-13377-y