Skip to main content
Log in

Elements for a general memory structure: properties of recurrent neural networks used to form situation models

  • Original Paper
  • Published:
Biological Cybernetics Aims and scope Submit manuscript

Abstract

We study how individual memory items are stored assuming that situations given in the environment can be represented in the form of synaptic-like couplings in recurrent neural networks. Previous numerical investigations have shown that specific architectures based on suppression or max units can successfully learn static or dynamic stimuli (situations). Here we provide a theoretical basis concerning the learning process convergence and the network response to a novel stimulus. We show that, besides learning ”simple“ static situations, a nD network can learn and replicate a sequence of up to n different vectors or frames. We find limits on the learning rate and show coupling matrices developing during training in different cases including expansion of the network into the case of nonlinear interunit coupling. Furthermore, we show that a specific coupling matrix provides low-pass-filter properties to the units, thus connecting networks constructed by static summation units with continuous-time networks. We also show under which conditions such networks can be used to perform arithmetic calculations by means of pattern completion.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  • Beer RD (2006) Parameter space structure of continuous-time recurrent neural networks. Neural Comput 18: 3009–3051

    Article  PubMed  Google Scholar 

  • Cruse H, Hübner D (2008) Selforganizing memory: active learning of landmarks used for navigation. Biol Cybern (submitted)

  • Cruse H, Sievers K (2008) A general network structure for learning Pavlovian paradigms (in preparation)

  • Elman JL (1990) Finding structure in time. Cogn Sci 14: 179–211

    Article  Google Scholar 

  • Feynman R (2001) In: Hawking SW (ed) The universe in a nutshell. Bantam Press, New York

  • Fuster JM (1995) Memory in the cerebral cortex: an empirical approach to neural networks in the human and nonhuman primate. MIT Press, Cambridge

    Google Scholar 

  • Hopfield JJ (1982) Neural networks and physical systems with emergent collective computational abilities. Proc Natl Acad Sci 79: 2554–2558

    Article  CAS  PubMed  Google Scholar 

  • Hopfield JJ (1984) Neurons with graded response have collective computational properties like those of two state neurons. Proc Natl Acad Sci 81: 3088–3092

    Article  CAS  PubMed  Google Scholar 

  • Jaeger H, Haas H (2004) Harnessing nonlinearity: predicting chaotic systems and saving energy in wireless communication. Science 2:78–80

    Article  Google Scholar 

  • Kühn S, Beyn WJ, Cruse H (2007) Modelling memory functions with recurrent neural networks consisting of input compensation units: I. Static situations. Biol Cybern 96: 455–470

    Article  PubMed  Google Scholar 

  • Kühn S, Cruse H (2007) Modelling memory functions with recurrent neural networks consisting of input compensation units: II. Dynamic situations. Biol Cybern 96: 471–486

    Article  PubMed  Google Scholar 

  • Kindermann T, Cruse H (2002) MMC— a new numerical approach to the kinematics of complex manipulators. Mech Mach Theory 37: 375–94

    Article  Google Scholar 

  • Palm G, Sommer FT (1996) Associative data storage and retrieval in neural networks. In: Domany E, van Hemmen JL, Schulten K(eds) Models of neural networks III. Association, generalization, and representation. Springer, New York, pp 79–18

    Google Scholar 

  • Pasemann F (2002) Complex dynamics and the structure of small neural networks. Netw: Comput Neural Syst 13: 195–16

    Google Scholar 

  • Steinkühler U, Cruse H (1998) A holistic model for an internal representation to control the movement of a manipulator with redundant degrees of freedom. Biol Cybern 79: 457–66

    Article  Google Scholar 

  • Strang G (2003) Introduction to linear algebra. Wellesley - Cambridge Press, Cambridge

    Google Scholar 

  • Tani J (2003) Learning to generate articulated behavior through the bottom-up and the top-down interaction processes. Neural Netw 16: 11–3

    Article  PubMed  Google Scholar 

  • Wessnitzer J, Webb B (2006) Multimodal sensory integration in insects—towards insect brain control architectures. Bioinspir Biomim 1: 63–5

    Article  PubMed  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Holk Cruse.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Makarov, V.A., Song, Y., Velarde, M.G. et al. Elements for a general memory structure: properties of recurrent neural networks used to form situation models. Biol Cybern 98, 371–395 (2008). https://doi.org/10.1007/s00422-008-0221-5

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00422-008-0221-5

Keywords