Skip to main content

Distant Co-occurrence Language Model for ASR in LooseWord Order Languages

  • Conference paper
Computer Recognition Systems 4

Part of the book series: Advances in Intelligent and Soft Computing ((AINSC,volume 95))

  • 1284 Accesses

Abstract

In the paper the problem of language modeling for automatic speech recognition in loose word order languages is considered. In loose word order languages classical n-gram language models are less effective, because the ordered word sequences encountered in the language corpus used to build the language models are less specific than in the case of strict word order languages. Because a word set appearing in the phrase is likely to appear in other permutation, all permutations of word sequences encountered in the corpus should be given additional likelihood in the language model.We propose the method of n-gram language model construction which assigns additional probability to word tuples being permutations of word sequences found in the training corpus. The paradigm of backoff bigram language model is adapted. The modification of typical model construction method consists in increasing the backed-off probability of bigrams that never appeared in the corpus but which elements appeared in the same phrases separated by other words. The proposed modification can be applied to any method of language model construction that is based on ML probability discounting. The performances of various LM creation methods adapted with the proposed way were compared in the application to Polish speech recognition.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 259.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 329.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Jelinek, F., Merialdo, B., Roukos, S., Strauss, M.: A dynamic language model for speech recognition. In: Proceedings of the Workshop on Speech and Natural Language, HLT 1991, Association for Computational Linguistics, pp. 293–295 (1991)

    Google Scholar 

  2. Piasecki, M., Broda, B.: Correction of medical handwriting OCR based on semantic similarity. In: Yin, H., Tino, P., Corchado, E., Byrne, W., Yao, X. (eds.) IDEAL 2007. LNCS, vol. 4881, pp. 437–446. Springer, Heidelberg (2007)

    Chapter  Google Scholar 

  3. Devine, E.G., Gaehde, S.A., Curtis, A.C.: Comparative Evaluation of Three Continuous Speech Recognition Software Packages in the Generation of Medical Reports. Journal of American Medical Informatics Association 7(5), 462–468 (2000)

    Article  Google Scholar 

  4. Chen, s.F., Goodman, S.: An empirical study of smoothing techniques for language modeling. Computer Speech and Language (13), 359–394 (1999)

    Google Scholar 

  5. Ziolko, B., Skurzok, D., Ziolko, M.: Word n-grams for Polish. In: Proc. of 10th IASTED Int. Conf. on Artifficial Intelligence and Applications (AIA 2010), pp. 197–201 (2010)

    Google Scholar 

  6. Mauces, M., Rotownik, T., Zemljak, M.: Modelling Highly Inflected Slovenian Language. International Journal of Speech technology 6, 254–257 (2003)

    Google Scholar 

  7. Whittaker, E.W.D., Woodland, P.C.: Language modelling for Russian and English using words and classes. In: Computer Speech and Language, vol. 17, pp. 87–104. Elsevier Academic Press, Amsterdam (2003)

    Google Scholar 

  8. Joshua, T., Goodman, J.T.: A Bit of Progress in Language Modeling Extended Version. Machine Learning and Applied Statistics Group Microsoft Research. Technical Report, MSR-TR-2001-72 (2001)

    Google Scholar 

  9. Katz, S.: Estimation of Probabilities from Sparse Data for the Language Model Component of a Speech Recognizer. IEEE Transactions On Acoustics, Speech and Signal Processing ASP-35(3), 400–401 (1987)

    Article  Google Scholar 

  10. Jurafsky, D., Matrin, J.: Speech and language processing. An introduction to natural language processing. In: Computational Linguistics and Speech Recognition, Pearson Prentice Hall, New Jersey (2009)

    Google Scholar 

  11. Gale, A., Sampson, G.: Good-Turing frequency estimation without tears. Journal of Quantitative Linguistics 2, 217–239 (1995)

    Article  Google Scholar 

  12. Lee, A., Kawahara, T., Shikano, K.: Julius - an Open Source Real-Time Large Vocabulary Recognition Engine. In: Proc. of European Conference on Speech Communication and Technology (EUROSPEECH), pp. 1691–1694 (2001)

    Google Scholar 

  13. Young, S., Everman, G.: The HTK Book (for HTK Version 3.4). Cambridge University Engineering Department (2009)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2011 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Sas, J., Zolnierek, A. (2011). Distant Co-occurrence Language Model for ASR in LooseWord Order Languages. In: Burduk, R., Kurzyński, M., Woźniak, M., Żołnierek, A. (eds) Computer Recognition Systems 4. Advances in Intelligent and Soft Computing, vol 95. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-20320-6_78

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-20320-6_78

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-20319-0

  • Online ISBN: 978-3-642-20320-6

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics