Abstract
Sleep is known to play a crucial role in memory and learning processes. During sleep, the replay of neural activity patterns associated with recent memories is believed to contribute to memory consolidation. Inspired by the sleep replay process in the biological system, we proposed a brain-like memory replay algorithm consisting of three phases. First, an artificial neural network (ANN) underwent learning, mimicking the cognitive processes of the awake stage. Next, we conducted a simulation where slow oscillation sleep or awake stages were simulated in a spiking neural network (SNN). The network connections and initial synaptic weights are identical to those of the trained ANN. Lastly, the performance of the ANN is evaluated and validated during the awake phase using the synaptic weights converted from the SNN. We applied this brain-inspired memory replay algorithm to a two-digit MNIST classification task. We found that as the level of noise increases, the model exhibits a significant improvement in performance after a period of sleep compared to the awake state. These findings provide compelling evidence that sleep-dependent memory consolidation can significantly enhance the network performance and improve its the robustness. Overall, this study sheds light on the potential advantages of incorporating sleep-like processes into neural networks, offering valuable insights into the field.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Abiodun, O.I., et al.: Comprehensive review of artificial neural network applications to pattern recognition. IEEE Access 7, 158820–158846 (2019)
Wu, Y., Feng, J.: Development and application of artificial neural network. Wirel. Pers. Commun. 102, 1645–1656 (2018)
Abiodun, O.I., Jantan, A., Omolara, A.E., Dada, K.V., Mohamed, N.A., Arshad, H.: State-of-the-art in artificial neural network applications: a survey. Heliyon 4, e00938 (2018)
Geirhos, R., Temme, C.R.M., Rauber, J., Schütt, H.H., Bethge, M., Wichmann, F.A.: Generalisation in humans and deep neural networks. http://arxiv.org/abs/1808.08750 (2020)
Hassabis, D., Kumaran, D., Summerfield, C., Botvinick, M.: Neuroscience-inspired artificial intelligence. Neuron 95, 245–258 (2017)
Rasch, B., Born, J.: About sleep’s role in memory. Physiol. Rev. 93, 681–766 (2013)
Frank, M.G.: Erasing synapses in sleep: is it time to be SHY? Neural Plast. 2012, 264378 (2012)
Achermann, P., BorbÉly, A.: Temporal evolution of coherence and power in the human sleep electroencephalogram. J. Sleep Res. 7, 36–41 (1998)
Girardeau, G., Lopes-dos-Santos, V.: Brain neural patterns and the memory function of sleep. Science 374, 560–564 (2021)
Wei, Y., Krishnan, G.P., Komarov, M., Bazhenov, M.: Differential roles of sleep spindles and sleep slow oscillations in memory consolidation. PLoS Comput. Biol. 14, e1006322 (2018)
Wei, Y., Krishnan, G.P., Marshall, L., Martinetz, T., Bazhenov, M.: Stimulation augments spike sequence replay and memory consolidation during slow-wave sleep. J. Neurosci. 40, 811–824 (2020)
Wei, Y., Krishnan, G.P., Bazhenov, M.: Synaptic mechanisms of memory consolidation during sleep slow oscillations. J. Neurosci. 36, 4231–4247 (2016)
Bellec, G., Salaj, D., Subramoney, A., Legenstein, R., Maass, W.: Long short-term memory and learning-to-learn in networks of spiking neurons. Presented at the 32nd Conference on Neural Information Processing Systems (NIPS 2018) (2018)
Krishnan, G.P., Tadros, T., Ramyaa, R., Bazhenov, M.: Biologically inspired sleep algorithm for artificial neural networks (2019). http://arxiv.org/abs/1908.02240
Lee, C., Panda, P., Srinivasan, G., Roy, K.: Training deep spiking convolutional neural networks with STDP-based unsupervised pre-training followed by supervised fine-tuning. Front. Neurosci. 12, 435 (2018)
Liu, F., Zhao, W., Chen, Y., Wang, Z., Yang, T., Jiang, L.: SSTDP: supervised spike timing dependent plasticity for efficient spiking neural network training. Front. Neurosci. 15 (2021)
Tadros, T., Krishnan, G.P., Ramyaa, R., Bazhenov, M.: Sleep-like unsupervised replay reduces catastrophic forgetting in artificial neural networks. Nat. Commun. 13, 7742 (2022)
Yi, Z., Lian, J., Liu, Q., Zhu, H., Liang, D., Liu, J.: Learning rules in spiking neural networks: a survey. Neurocomputing 531, 163–179 (2023)
Yu, Q., Tang, H., Tan, K.C., Yu, H.: A brain-inspired spiking neural network model with temporal encoding and learning. Neurocomputing 138, 3–13 (2014)
Zhang, T., Zeng, Y., Zhao, D., Xu, B.: Brain-inspired balanced tuning for spiking neural networks. In: Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, Stockholm, Sweden, pp. 1653–1659. International Joint Conferences on Artificial Intelligence Organization (2018)
Chen, G., Scherr, F., Maass, W.: A data-based large-scale model for primary visual cortex enables brain-like robust and versatile visual processing. Sci. Adv. 8, eabq7592 (2022)
Singh, D., Norman, K.A., Schapiro, A.C.: A model of autonomous interactions between hippocampus and neocortex driving sleep-dependent memory consolidation. Proc. Natl. Acad. Sci. U.S.A. 119, e2123432119 (2022)
Marković, D., Mizrahi, A., Querlioz, D., Grollier, J.: Physics for neuromorphic computing. Nat. Rev. Phys. 2, 499–510 (2020)
Pei, J., et al.: Towards artificial general intelligence with hybrid Tianjic chip architecture. Nature 572, 106–111 (2019)
Wu, Y., et al.: Brain-inspired global-local learning incorporated with neuromorphic computing. Nat. Commun. 13, 65 (2022)
Zenke, F., Neftci, E.O.: Brain-inspired learning on neuromorphic substrates. Proc. IEEE 109, 935–950 (2021)
Zhang, Y., et al.: A system hierarchy for brain-inspired computing. Nature 586, 378–384 (2020)
Zhao, R., et al.: A framework for the general design and computation of hybrid neural networks. Nat. Commun. 13, 3427 (2022)
Nir, Y., et al.: Regional slow waves and spindles in human sleep. Neuron 70, 153–169 (2011)
Miyawaki, H., Watson, B.O., Diba, K.: Neuronal firing rates diverge during REM and homogenize during non-REM. Sci. Rep. 9, 689 (2019)
Bi, G., Poo, M.: Synaptic modifications in cultured hippocampal neurons: dependence on spike timing, synaptic strength, and postsynaptic cell type. J. Neurosci. 18, 10464–10472 (1998)
Park, Y., Choi, W., Paik, S.-B.: Symmetry of learning rate in synaptic plasticity modulates formation of flexible and stable memories. Sci. Rep. 7, 5671 (2017)
Froemke, R.C., Dan, Y.: Spike-timing-dependent synaptic modification induced by natural spike trains. Nature 416, 433–438 (2002)
Klinzing, J.G., Niethard, N., Born, J.: Mechanisms of systems memory consolidation during sleep. Nat. Neurosci. 22, 1598–1610 (2019)
Acknowledgement
This work was supported by the National Natural Science Foundation of China (12101570), Zhejiang Lab & Pujiang Lab (K2023KA1BB01), and Scientific Projects of Zhejiang Lab (K2023KI0AA02, 2021KE0PI03).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Xie, S., Tang, T., Feng, L., Lin, F., Wei, Y. (2023). Sleep-Dependent Memory Replay Enables Brain-Like Robustness in Neural Networks. In: Yang, H., et al. Intelligent Robotics and Applications. ICIRA 2023. Lecture Notes in Computer Science(), vol 14267. Springer, Singapore. https://doi.org/10.1007/978-981-99-6483-3_19
Download citation
DOI: https://doi.org/10.1007/978-981-99-6483-3_19
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-6482-6
Online ISBN: 978-981-99-6483-3
eBook Packages: Computer ScienceComputer Science (R0)