Skip to main content

Ensemble Methods for Word Embedding Model Based on Judicial Text

  • Conference paper
  • First Online:
Web Information Systems and Applications (WISA 2019)

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 11817))

Included in the following conference series:

Abstract

With the continuous expansion of computer applications, scenarios such as machine translation, speech recognition, and message retrieval depend on the techniques of the natural language processing. As a technique for training word vectors, Word2vec is widely used because it can train word embedding model based on corpus and represent the sentences as vectors according to the training model. However, as an unsupervised learning model, word embedding can only characterize the internal relevance of natural language in non-specific scenarios. For a specific field like judicial, the method of expanding the vector space by creating a professional judicial corpus to enhance the accuracy of similarity calculation is not obvious, and this method is unable to provide further analysis for similarity in cases belonging to the same type. Therefore, based on the original word embedding model, we extract factors such as fines and prison term to help identify the differences, and attach the label of the case to complete supervised ensemble learning. The result of the ensemble model is better than any result of single model in terms of distinguishing whether they are the same type. The experimental result also reveal that the ensemble method can effectively tell the difference between similar cases, and is less sensitive to the details of the training data, the choice of training plan and the contingency of a single inaccurate training run.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Al-Kofahi, K., Jackson, P., Travers, T.E., Tyrell, A.: Systems, methods, and software for classifying text from judicial opinions and other documents, uS Patent 7,062,498, June 2006

    Google Scholar 

  2. Barlow, H.B.: Unsupervised learning. Neural Comput. 1(3), 295–311 (1989)

    Article  Google Scholar 

  3. Deng, L., Platt, J.C.: Ensemble deep learning for speech recognition. In: Fifteenth Annual Conference of the International Speech Communication Association (2014)

    Google Scholar 

  4. Grbovic, M., Cheng, H.: Real-time personalization using embeddings for search ranking at Airbnb. In: Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pp. 311–320. ACM (2018)

    Google Scholar 

  5. He, T., Lian, H., Qin, Z., Zou, Z., Luo, B.: Word embedding based document similarity for the inferring of penalty. In: Meng, X., Li, R., Wang, K., Niu, B., Wang, X., Zhao, G. (eds.) WISA 2018. LNCS, vol. 11242, pp. 240–251. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-02934-0_22

    Chapter  Google Scholar 

  6. Hinton, G., et al.: Deep neural networks for acoustic modeling in speech recognition. IEEE Signal Process. Magaz. 29, 82 (2012)

    Article  Google Scholar 

  7. Iwayama, M., Tokunaga, T.: Cluster-based text categorization: a comparison of category search strategies. In: Proceedings of the 18th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 273–280. Citeseer (1995)

    Google Scholar 

  8. Jaderberg, M., Vedaldi, A., Zisserman, A.: Deep features for text spotting. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8692, pp. 512–528. Springer, Cham (2014). https://doi.org/10.1007/978-3-319-10593-2_34

    Chapter  Google Scholar 

  9. Jain, A.K., Farrokhnia, F.: Unsupervised texture segmentation using gabor filters. Pattern Recognit. 24(12), 1167–1186 (1991)

    Article  Google Scholar 

  10. Levy, O., Goldberg, Y.: Neural word embedding as implicit matrix factorization. In: Advances in Neural Information Processing Systems, pp. 2177–2185 (2014)

    Google Scholar 

  11. Marti, U.V., Bunke, H.: Using a statistical language model to improve the performance of an hmm-based cursive handwriting recognition system. In: Hidden Markov Models: Applications in Computer Vision, pp. 65–90. World Scientific (2001)

    Google Scholar 

  12. Mikolov, T., Chen, K., Corrado, G., Dean, J.: Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781 (2013)

  13. Tang, D., Wei, F., Yang, N., Zhou, M., Liu, T., Qin, B.: Learning sentiment-specific word embedding for twitter sentiment classification. In: Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), vol. 1, pp. 1555–1565 (2014)

    Google Scholar 

  14. Turian, J., Ratinov, L., Bengio, Y.: Word representations: a simple and general method for semi-supervised learning. In: Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics, pp. 384–394. Association for Computational Linguistics (2010)

    Google Scholar 

  15. Zhang, C., Ma, Y.: Ensemble Machine Learning: Methods and Applications, 1st edn. Springer, New York (2012). https://doi.org/10.1007/978-1-4419-9326-7

    Book  MATH  Google Scholar 

Download references

Acknowledgment

The work is supported in part by the National Key Research and Development Program of China (2016YFC0800805) and the National Natural Science Foundation of China (61772014).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Tieke He .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2019 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Xia, C., He, T., Wan, J., Wang, H. (2019). Ensemble Methods for Word Embedding Model Based on Judicial Text. In: Ni, W., Wang, X., Song, W., Li, Y. (eds) Web Information Systems and Applications. WISA 2019. Lecture Notes in Computer Science(), vol 11817. Springer, Cham. https://doi.org/10.1007/978-3-030-30952-7_31

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-30952-7_31

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-30951-0

  • Online ISBN: 978-3-030-30952-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics