Skip to main content

Modeling Primacy, Recency, and Cued Recall in Serial Memory Task Using On-Center Off-Surround Recurrent Neural Network

  • Conference paper
  • First Online:
Machine Learning, Optimization, and Data Science (LOD 2023)

Abstract

The serial recall paradigm has long been used to study short-term memory. Previous experiments have consistently revealed two key phenomena: the primacy effect and the recency effect. Essentially, it is easier to recall items at the beginning and end of a sequence compared to those in the middle. In this study, we present a single-layer fully connected recurrent neural network with self-excitation and mutual inhibition. By providing transient input to the network, we observed a dynamic steady-state output. We examined this output pattern to determine which inputs were ultimately “remembered” at the end of the simulation. Our results demonstrate that this network can replicate the serial recall curve observed in empirical studies when higher inhibition values are used. Additionally, it can account for the capacity limitation commonly observed in serial recall tasks. By varying the presentation duration in the model, we successfully explain both the primacy and recency effects within the same network. Furthermore, when we introduced cues to a single item in the sequence by elevating its input, we observed a decrease in the recall probability of neighboring items. In summary, our findings suggest that the dynamics of a single-layer recurrent on-center off-surround neural network can provide insights into the mechanisms underlying primacy, recency, and cued recall effects.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 69.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 89.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Baddeley, A.: Working memory: looking back and looking forward. Nat. Rev. Neurosci. 4(10), 829–839 (2003)

    Article  Google Scholar 

  2. Miller, A.G.: The magical number seven, plus or minus two: some limits on our capacity for processing information. Psychol. Rev. 63, 81–97 (1956)

    Article  Google Scholar 

  3. Sreenivasan, K.K., Curtis, C.E., D’Esposito, M.: Revisiting the role of persistent neural activity during working memory. Trends Cogn. Sci. 18(2), 82–89 (2014)

    Article  Google Scholar 

  4. Goldman-Rakic, P.: Cellular basis of working memory. Neuron 14(3), 477–485 (1995)

    Article  Google Scholar 

  5. Baddeley, A.: Working memory: theories, models, and controversies. Annu. Rev. Psychol. 63, 1–29 (2012)

    Article  Google Scholar 

  6. Cowan, N.: The magical number 4 in short-term memory: a reconsideration of mental storage capacity. Behav. Brain Sci. 24, 87–185 (2001)

    Article  Google Scholar 

  7. Luck, S.J., Vogel, E.K.: The capacity of visual working memory for features and conjunctions. Nature 390, 279 (1997)

    Article  Google Scholar 

  8. Zhang, W., Luck, S.J.: Discrete fixed-resolution representations in visual working memory. Nature 453(7192), 233–235 (2008)

    Article  Google Scholar 

  9. Ma, W.J., Husain, M., Bays, P.M.: Changing concepts of working memory. Nat. Neurosci. 17(3), 347–356 (2014)

    Article  Google Scholar 

  10. Waugh, N.C., Norman, D.A.: Primary memory. Psychol. Rev. 72, 89–104 (1965)

    Article  Google Scholar 

  11. van den Berg, R., Shin, H., Chou, W., George, R., Ma, W.J.: Variability in encoding precision accounts for visual short-term memory limitations. Proc. Natl. Acad. Sci. 109(22), 8780–8785 (2012)

    Article  Google Scholar 

  12. Usher, M., Cohen, J.D.: Short term memory and selection processes in a frontal-lobe model. In: Heinke, D., Humphreys, G.W., Olson, A. (eds.) Connectionist Models in Cognitive Neuroscience, pp. 78–91 (1999)

    Google Scholar 

  13. Grossberg, S.: Contour enhancement, short term memory, and constancies in reverberating neural networks. In: Studies in Applied Mathematics, vol. LII, pp. 213–257 (1973)

    Google Scholar 

  14. Sengupta, R., Surampudi, B.R., Melcher, D.: A visual sense of number emerges from the dynamics of a recurrent on-center off-surround neural network. Brain Res. 1582, 114–224 (2014)

    Article  Google Scholar 

  15. Knops, A., Piazza, M., Sengupta, R., Eger, E., Melcher, D.: A shared, flexible neural map architecture reflects capacity limits in both visual short term memory and enumeration. J. Neurosci. 34(30), 9857–9866 (2014)

    Article  Google Scholar 

  16. Tsotsos, J.K., Kruijne, W.: Cognitive programs: software for attention’s executive. Front. Psychol. 5, 1260 (2014)

    Article  Google Scholar 

  17. Masse, N.Y., Yang, G.R., Song, H.F., Wang, X.J., Freedman, D.J.: Circuit mechanisms for the maintenance and manipulation of information in working memory. Nat. Neurosci. 22(7), 1159–1167 (2019)

    Article  Google Scholar 

  18. Huang, Q.-S., Wei, H.: A computational model of working memory based on spike-timing-dependent plasticity. Front. Comput. Neurosci. 15, 630999 (2021)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Rakesh Sengupta .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Vindhya, L.S., Gnana Prasanna, R., Sengupta, R., Shukla, A. (2024). Modeling Primacy, Recency, and Cued Recall in Serial Memory Task Using On-Center Off-Surround Recurrent Neural Network. In: Nicosia, G., Ojha, V., La Malfa, E., La Malfa, G., Pardalos, P.M., Umeton, R. (eds) Machine Learning, Optimization, and Data Science. LOD 2023. Lecture Notes in Computer Science, vol 14505. Springer, Cham. https://doi.org/10.1007/978-3-031-53969-5_30

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-53969-5_30

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-53968-8

  • Online ISBN: 978-3-031-53969-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics