Elsevier

Neural Networks

Volume 15, Issue 3, April 2002, Pages 327-335
Neural Networks

Contributed araticle
A local and neurobiologically plausible method of learning correlated patterns

https://doi.org/10.1016/S0893-6080(02)00031-XGet rights and content

Abstract

The problem of learning correlated patterns in a neurobiologically plausible way without multiple iterations is discussed. Three guiding principles are outlined for solving this problem in a manner suggestive of how the memory modules of the human brain do it. The first, already known, principle of minimum disturbance is applied quantitatively to minimise the number of learning iterations. The second guiding principle involves a self-induced remedy to undo any damage caused to the stabilities of the old stored patterns after a new one has been learnt. The third involves localising connectivities between neurons depending on the structure of the input patterns. A neurobiologically plausible network and a learning rule are constructed based on these principles. Satisfactory use of this network in learning English words as sequences of their letters is demonstrated by means of computer simulation.

Introduction

Modelling the pattern learning capabilities of the human brain is an active area of research in recent times. It is now generally understood that patterns are stored in the synapses between the cortical neurons in the form of connection strengths and each pattern corresponds to a stable or fixed mode of activities of the neurons. The finite basin of attraction around a fixed mode of activity is supposed to provide the required noise tolerance in recalling an intended pattern. Based on a model proposed by Hopfield (1982), much work has been done subsequently (van Hemmen and Kuhn, 1991, Hertz et al., 1991) in trying to understand the storage of patterns in the human brain. Many Hebbian and local rules have been proposed for learning extensively many random and uncorrelated patterns. Both analytical and numerical studies have shown their efficacy (Amit et al., 1987, Amit, 1989) beyond doubt. However, when the patterns to be stored are correlated, most known rules are unable to learn them in just one presentation. Invariably many iterations of learning are required before all the patterns are stored as fixed modes of activity of the network (Krauth and Mezard, 1987, Gardner, 1988). With only one presentation most correlated patterns are not stored correctly; moreover, undesirable linear combinations of them get stored as spurious patterns.

The Hopfield model and many of its variants deal with high activity patterns. Experimental studies in neurobiology suggest that the average activity of any module of the brain is very low (Abeles, 1982), certainly well below the 50% assumed in the Hopfield model. Taking this into account Tsodyks and Feigel'man proposed a model for learning low activity patterns (Tsodyks & Feigel'man, 1988). Their model uses 0/1 coding for representing passive and active neurons. Its storage capacity turns out to be far greater than that of the Hopfield model when the activity levels of the patterns to be stored are well below 50%. However, it also cannot learn correlated patterns; specifically, patterns having correlations among their active elements. The same is true of a model proposed by Willshaw et al. (Golomb et al., 1990, Willshaw et al., 1969).

It is, of course, straightforward to construct an iterative learning rule requiring multiple presentations of low activity and correlated patterns. Locality or Hebbian condition can also be enforced. However, an external supervisory mechanism is needed to check if all the patterns are correctly learnt and if the iterations can stop. Patterns encountered in practical life are seldom uncorrelated with each other and a biologically plausible learning mechanism must be able to do with one presentation of a pattern to be learnt. Often, a second presentation of the pattern may not be feasible except in ideal classroom situations. Many times we are able to recall a pattern, say a face, on being presented with a clue long after learning it for the first time. In the intervening period we might not have had a chance to explicitly learn it again nor might we have learnt anything associated with it. Moreover, we might have learnt other patterns of a similar nature. In spite of all this, we do recall an old pattern, learnt only once, correctly when we need to. For practical applications too a rule capable of learning patterns in a single presentation would be useful. Learning, then, would become automatic and extra components on the learning system such as a teacher and pattern lookup storage, etc. could be dispensed with.

One of the reasons why most learning rules call for multiple presentations of correlated patterns is that while trying to ensure the correct level of stability of a new pattern, they either undercorrect or overcorrect the synaptic strengths. If the corrections to the synaptic strengths were less than what are required, obviously multiple presentations of the new pattern would become necessary. On the other hand, if the corrections are overdone, they are bound to disturb the stability of some old stored pattern. As a result, that old pattern would have to be relearnt. Correlated patterns are more likely to disturb the stabilities of each other than uncorrelated ones. Therefore, the corrections to the synaptic strengths have to be carefully worked out. As part of a numerical study being reported in this paper a learning rule, which can incorporate the right amount of corrections was worked out and experimented. It has been found that additional principles and mechanisms are needed if correlated patterns have to be learnt in one presentation. One of them is a method of self-induced relearning to cure the damages caused to the stabilities of the old patterns when learning a new one. Another is to localise the connectivities of the network in such a way as to suit the structure of the patterns to be stored. A neural network model and a learning rule based on these principles are described in the following sections. Experimental verification of the efficiency of the network is provided by means of learning sets of English words in one presentation.

The paper is organised as follows. Firstly, the three guiding principles are presented one of which has already been proposed in the literature. Based on these principles a new network for associative memory and a learning rule are constructed. Experimental verification of the performance of this network in learning correlated patterns in one presentation is presented next. The paper concludes with a discussion on the generality of the guiding principles for constructing biologically plausible associative memory networks.

Section snippets

Three guiding principles

The first guiding principle is called the principle of minimum disturbance and is clearly formulated by Widrow and Lehr (1990) in their review. The main point of their study is that when learning to classify a new pattern, only absolutely minimum changes must be incorporated into the network so as to ensure a correct classification. They also demonstrate the application of this principle on a number of pattern classifying networks namely, adaline and its variants. In most cases the application

A new associative memory network

The key issue in evolving a new network model for learning correlated patterns is to see whether the network would facilitate a quantitative application of the principle of minimum disturbance. In seeking to do that, it has been found that adherence to neurobiological constraints has proven to be a source of help rather than a problem. In living brains an excitatory synapse does not change into an inhibitory one. Both excitatory and inhibitory learning take place separately and concurrently.

A sample application of learning words

Some of the most frequently used patterns are words in any natural language. Seen as sequences of their letters they are highly correlated and are difficult to learn using most outer-product type of Hebbian rules. While written words are represented by sequences of letters, phonemes constitute spoken words. Both written and spoken words are arguably stored and recalled as static patterns though their constituents stream into the human brain through the respective sense organs in time. It is

Conclusion

Some of the interesting architectural features of the new network are the separate variables to represent the excitatory and inhibitory synapses and the connection frequency matrix. Learning in excitatory and inhibitory synapses takes place concurrently but distinctly as neurobiological evidence suggests (Abrams, 1994). Bounds on the growth of their strengths can explain saturation of the network's ability to learn. In the context of human brain it is not known whether nonstop learning is ever

References (20)

  • D.J. Amit et al.

    Statistical mechanics of neural networks near saturation

    Annals of Physics (New York)

    (1987)
  • M. Abeles

    Local cortical circuits

    (1982)
  • T.W. Abrams

    Learning and memory cellular studies in simple neural systems

  • D.J. Amit

    Modeling brain functions

    (1989)
  • Athithan, G. (1997). Efficient modelling and performance simulation of neural associative memory networks. PhD thesis....
  • G. Athithan

    Associative memory of low activity patterns without the problem of spurious patterns

    Current Science

    (1999)
  • G. Athithan et al.

    On the problem of spurious patterns in neural associative memory models

    IEEE Transactions on Neural Networks

    (1997)
  • F. Crick et al.

    The function of dream sleep

    Nature

    (1983)
  • E. Gardner

    The space of interactions in neural network models

    Journal of Physics A: Mathematics and General

    (1988)
  • D. Golomb et al.

    Willshaw model—associative memory with sparse coding and low firing rates

    Physical Review A

    (1990)
There are more references available in the full text version of this article.

Cited by (0)

View full text