Skip to main content

Applying Model Fusion to Augment Data for Entity Recognition in Legal Documents

  • Conference paper
  • First Online:
Natural Language Processing and Chinese Computing (NLPCC 2020)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 12430))

  • 3083 Accesses

Abstract

Named entity recognition for legal documents is a basic and crucial task, which can provide important knowledge for the related tasks in the field of wisdom justice. However, it is still difficult to augment the labeled data of named entities for legal documents automatically. To address this issue, we propose a novel data augmentation method for named entity recognition by fusing multiple models. Firstly, we train a total of ten models by conducting 5-fold cross-training on the small-scale labeled datasets based on Bilstm-CRF and Bert-Bilstm-CRF models separately. Next, we try to apply single-model fusion and multi-model fusion modes, in which, single-model fusion is to vote on the prediction results of five models of the same baseline, while multi-model fusion is to vote on the prediction results of ten models with two different baselines. Further, we take the identified entities with high correctness in the multiple experimental results as effective entities, and add them to the training set for the next training. Finally, we conduct the different experiments on two public datasets and our built judicial dataset separately, which shows the experimental results using data augmentation are close to those based on 5 times of labeled dataset, and obviously better than those on the initial small-scale labeled datasets.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Luo, B., Feng, Y., Xu, J., Zhao, D.: Learning to predict charges for criminal cases with legal basis. In: Empirical Methods in Natural Language Processing, pp. 2727–2736 (2017)

    Google Scholar 

  2. Jiang, X., Ye, H., Luo, Z., Chao, W., Ma, W.: Interpretable rationale augmented charge prediction system. In: International Conference on Computational Linguistics, pp. 146–151 (2018)

    Google Scholar 

  3. Lauderdale, B.E., Clark, T.S.: The supreme court’s many median justices. Am. Polit. Sci. Rev. 106(04), 847–866 (2012)

    Article  Google Scholar 

  4. Duan, X., et al.: CJRC: a reliable human-annotated benchmark dataset for chinese judicial reading comprehension. In: Sun, M., Huang, X., Ji, H., Liu, Z., Liu, Y. (eds.) CCL 2019. LNCS (LNAI), vol. 11856, pp. 439–451. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-32381-3_36

    Chapter  Google Scholar 

  5. Wang, L., Yan, Q., Li, S., Zhou, G.: Employing auto-annotated data for person name recognition in judgment documents. In: Sun, M., Wang, X., Chang, B., Xiong, D. (eds.) CCL/NLP-NABD -2017. LNCS (LNAI), vol. 10565, pp. 13–23. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-69005-6_2

    Chapter  Google Scholar 

  6. Cardellino, C., Teruel, M., Alemany, L.A., Villata, S.: A low-cost, high-coverage legal named entity recognizer, classifier and linker. In: 16th International Conference on Artificial Intelligence and Law, pp. 9–18. Londres, United Kingdom (2017)

    Google Scholar 

  7. Leitner, E., Rehm, G., Morenoschneider, J.: Fine-grained named entity recognition in legal documents. In: 15th International Conference on Semantic Systems, pp. 272–287. Karlsruhe, Germany (2019)

    Google Scholar 

  8. Collobert, R., Weston, J., Bottou, L., Karlen, M., Kavukcuoglu, K., Kuksa, P.P.: Natural language processing (Almost) from scratch. J. Mach. Learn. Res. 12(1), 2493–2537 (2011)

    MATH  Google Scholar 

  9. Huang, Z., Xu, W., Yu, K.: Bidirectional LSTM-CRF Models for Sequence Tagging. arXiv preprint arXiv: 1508.01991 (2015)

    Google Scholar 

  10. Chiu, J.P., Nichols, E.: Named entity recognition with bidirectional LSTM-CNNs. Trans. Assoc. Comput. Linguist. 4(1), 357–370 (2016)

    Article  Google Scholar 

  11. Ma, X., Hovy, E.: End-to-end sequence labeling via bi-directional LSTM-CNNs-CRF. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, pp. 1064–1074 (2016)

    Google Scholar 

  12. Rei, M., Crichton, G.K., Pyysalo, S.: Attending to characters in neural sequence labeling models. In: International Conference on Computational Linguistics, pp. 309–318 (2016)

    Google Scholar 

  13. Bharadwaj, A., Mortensen, D.R., Dyer, C., Carbonellm, J.G.: Phonologically aware neural model for named entity recognition in low resource transfer settings. In: Empirical Methods in Natural Language Processing, pp. 1462–1472 (2016)

    Google Scholar 

  14. Tan, Z., Wang, M., Xie, J.: Deep Semantic Role Labeling with Self-Attention. arXiv preprint arXiv:1712.01586 (2017)

  15. Cetoli, A., Bragaglia, S., Oharney, A.D., Sloan, M.: Graph Convolutional Networks for Named Entity Recognition. arXiv preprint arXiv:1709.10053 (2017)

  16. Zhang, Y., Yang, J.: Chinese NER using lattice LSTM. In: Meeting of the Association for Computational Linguistics, pp. 1554–1564 (2018)

    Google Scholar 

  17. Peters, M.E., et al.: Deep contextualized word representations. In: North American Chapter of the Association for Computational Linguistics, pp. 2227–2237 (2018)

    Google Scholar 

  18. Devlin, J., Chang, M., Lee, K., Toutanova K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: North American Chapter of the Association for Computational Linguistics, pp. 4171–4186 (2018)

    Google Scholar 

  19. Lan, Z.: ALBERT: A Lite BERT for Self-supervised Learning of Language Representations. arXiv preprint arXiv: 1909.11942 https://arxiv.org/abs/1810.04805 (2019)

  20. Li, X., Zhang, H., Zhou, X.-H.: Chinese clinical named entity recognition with variant neural structures based on BERT Methods. J. Biomed. Inform. 107 (2020). https://doi.org/10.1016/j.jbi.2020.103422

  21. Moon, T., Awasthy, P., Ni, J., Florian, R.: Towards Lingua Franca Named Entity Recognition with BERT. arXiv preprint arXiv:1912.01389 (2019)

  22. Yu, A.W., et al.: QANet: Combining Local Convolution with Global Self-Attention for Reading Comprehension. arXiv preprint arXiv:1804.09541 (2018)

  23. Kobayashi, S.: Contextual augmentation: data augmentation by words with paradigmatic relations. North Am. Chapter Assoc. Comput. Linguist. 2, 452–457 (2018)

    Google Scholar 

  24. Samanta, S., Mehta, S.: Towards Crafting Text Adversarial Samples. arXiv preprint arXiv:1707.02812 (2017)

  25. Wu, D., Lee, W.S., Ye, N., Chieu, H.L.: Domain adaptive bootstrapping for named entity recognition. In: Empirical Methods in Natural Language Processing, pp. 1523–1532. Singapore (2009)

    Google Scholar 

  26. Neelakantan, A., Collins, M.: Learning dictionaries for named entity recognition using minimal supervision. In: Conference of the European Chapter of the Association for Computational Linguistics, pp. 452–461 (2014)

    Google Scholar 

  27. Peters, M.E., Ammar, W., Bhagavatula, C., Power, R.: Semi-supervised sequence tagging with bidirectional language models. Meet. Assoc. Comput. Linguist. 1, 1756–1765 (2017)

    Google Scholar 

Download references

Acknowledgments

This research was supported by the National Social Science Fund of China (No. 18BYY074).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Hu Zhang .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Zhang, H., Gao, H., Zhou, J., Li, R. (2020). Applying Model Fusion to Augment Data for Entity Recognition in Legal Documents. In: Zhu, X., Zhang, M., Hong, Y., He, R. (eds) Natural Language Processing and Chinese Computing. NLPCC 2020. Lecture Notes in Computer Science(), vol 12430. Springer, Cham. https://doi.org/10.1007/978-3-030-60450-9_20

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-60450-9_20

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-60449-3

  • Online ISBN: 978-3-030-60450-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics