Skip to main content

Appropriate Context Association and Learning Parameters for Word Spotting with Partially Recurrent Neural Networks

  • Conference paper
Neural Networks: Artificial Intelligence and Industrial Applications

Abstract

This paper covers part of a study which concerns the feasibility of real-time word spotting with partially recurrent neural networks (PRNN’s). PRNN’s have already proven appropriate for other examples of pure sequence recognition [1, 2]. However choices concerning architectural and learning aspects are still hard to make. One of the questions still to be answered, is how these aspects influence the term of memory of a PRNN. This paper tries to obtain some directives regarding architectures and learning algorithms.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Hertz J., Krogh A. and Palmer R.G., Introduction to the theory of neural computation, Addison-Wesley Publishing Company, Amsterdam, 1991.

    Google Scholar 

  2. Couwenberg R., Speech Recognition with Recurrent Neural Nets, University of Twente, Department of Electrical Engineering, Enschede, 1990.

    Google Scholar 

  3. Elman J.L., “Finding structure in time”, Cognitive Science, 14: 179–211, 1990.

    Article  Google Scholar 

  4. Minsky M. and Papert S., Perceptrons, Maple Press Company, Cambridge, MA, 1969.

    MATH  Google Scholar 

  5. Hinton G.E., Rumelhart D.E., and Williams R.J., “Learning Internal Representations by Error Propagation”, Parallel Distributed Processing, vol. 1, chap. 8, 1986.

    Google Scholar 

  6. Zipser D., “Subgrouping reduces complexity and speeds up learning in recurrent networks”, Advances in Neural Information Processing Systems II, pages 638–641, San Mateo, California, 1990.

    Google Scholar 

  7. Marslen-Wilson W.D., “Speech understanding as a psychological process”, J.C. Simon (ed.), Spoken language generation, Dordrecht: Riedel, 1980..

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 1995 Springer-Verlag London Limited

About this paper

Cite this paper

v. Leeuwen, D., Wittenburg, P., Poel, M. (1995). Appropriate Context Association and Learning Parameters for Word Spotting with Partially Recurrent Neural Networks. In: Kappen, B., Gielen, S. (eds) Neural Networks: Artificial Intelligence and Industrial Applications. Springer, London. https://doi.org/10.1007/978-1-4471-3087-1_48

Download citation

  • DOI: https://doi.org/10.1007/978-1-4471-3087-1_48

  • Publisher Name: Springer, London

  • Print ISBN: 978-3-540-19992-2

  • Online ISBN: 978-1-4471-3087-1

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics