Skip to main content

A Recurrent Neural Network Language Model Based on Word Embedding

  • Conference paper
  • First Online:
Web and Big Data (APWeb-WAIM 2018)

Abstract

Language model is one of the basic research issues of natural language processing, and which is the premise for realizing more complicated tasks such as speech recognition, machine translation and question answering system. In recent years, neural network language model has become a research hotspot, which greatly enhances the application effect of language model. In this paper, a recurrent neural network language model (RNNLM) based on word embedding is proposed, and the word embedding of each word is generated by pre-training the text data with skip-gram model. The n-gram language model, RNNLM based on one-hot and RNNLM based on word embedding are evaluated on three different public datasets. The experimental results show that the RNNLM based on word embedding performs best, and which can reduce the perplexity of language model significantly.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Jun, W.: The Beauty of Mathematics. Posts and Telecommunications Press, Beijing (2012)

    Google Scholar 

  2. Brown, P.F., Desouza, P.V., Mercer, R.L., Pietra, V.J.D., Lai, J.C.: Class-based n-gram models of natural language. Comput. Linguist. 18(4), 467–479 (1992)

    Google Scholar 

  3. Bengio, Y., Ducharme, R., Vincent, P., Jauvin, C.: A neural probabilistic language model. J. Mach. Learn. Res. 3(2), 1137–1155 (2003)

    MATH  Google Scholar 

  4. Mikolov, T., Karafiát, M., Burget, L., Černocký, J., Khudanpur, S.: Recurrent neural network based language model. In: Proceedings of the Eleventh Annual Conference of the International Speech Communication Association, pp. 1045–1048. International Speech Communication Association, Makuhari, Chiba (2010)

    Google Scholar 

  5. Bird, S., Loper, E.: NLTK: the natural language toolkit. In: Proceedings of the ACL 2004 on Interactive Poster and Demonstration Sessions, p. 31. Association for Computational Linguistics (2004)

    Google Scholar 

  6. Hinton, G.E.: Learning distributed representations of concepts. In: Proceedings of the Eighth Annual Conference of the Cognitive Science Society, vol. 1, p. 12. Oxford University Press, Oxford, UK (1986)

    Google Scholar 

  7. Mikolov, T., Chen, K., Corrado, G., Dean, J.: Efficient estimation of word representations in vector space. ArXiv Preprint arXiv:1301.3781 (2013)

  8. Theano 0.9.0 documentation. http://deeplearning.net/software/theano/. Accessed 10 Dec 2017

  9. Jian, Z.: Research on recurrent neural network language model in continuous speech recognition. The PLA Information Engineering University, Zhengzhou, pp. 19–20 (2014)

    Google Scholar 

  10. Recurrent neural networks tutorial. http://www.wildml.com/2015/09/recurrent-neural-networks-tutorial-part-2-implementing-a-language-model-rnn-with-python-numpy-and-theano/. Accessed 12 Jan 2018

  11. Marcus, M.P., Marcinkiewicz, M.A., Santorini, B.: Building a large annotated corpus of English: the Penn Treebank. Comput. Linguist. 19(2), 313–330 (1993)

    Google Scholar 

  12. Stolcke, A.: SRILM-an extensible language modeling toolkit. In: Proceedings of the Seventh International Conference on Spoken Language Processing, pp. 901–904 (2002)

    Google Scholar 

Download references

Acknowledgements

This work is supported by the Beijing Natural Science Foundation under Grant No. 4162067.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Shuaimin Li .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2018 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Li, S., Xu, J. (2018). A Recurrent Neural Network Language Model Based on Word Embedding. In: U, L., Xie, H. (eds) Web and Big Data. APWeb-WAIM 2018. Lecture Notes in Computer Science(), vol 11268. Springer, Cham. https://doi.org/10.1007/978-3-030-01298-4_30

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-01298-4_30

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-01297-7

  • Online ISBN: 978-3-030-01298-4

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics