Skip to main content
Log in

Spatiotemporal Backpropagation based on Channel Reward for Training High-Precision Spiking Neural Network

  • Original Paper
  • Published:
Signal, Image and Video Processing Aims and scope Submit manuscript

Abstract

Spiking neural network (SNN) has attracted much attention due to its spatial–temporal information processing ability and high biological reliability. SNN is a hardware-friendly model based on event-driven and sparse triggering, which can achieve low power consumption and efficient information processing on neural morphological chips. However, SNN processes discrete spike trains, and its unique working mode makes it more difficult to train than traditional networks. Based on this, this paper proposes a spatiotemporal backpropagation (STBP) algorithm for directly training high-performance SNN. By narrowing the coding time window, we convert the leaky integrate-and-fire (LIF) into its explicit iterative version for direct SNN training. To solve the problem of non-differentiability of SNN discrete spikes, we introduce an approximate derivative of spike firing for gradient descent training. We propose a biologically reasonable channel reward (CR) mechanism integrated into the STBP algorithm. This method makes full use of the spatial domain (SD) and time domain (TD) information of input and does not require any additional complex techniques. We evaluate the proposed algorithm on traditional static MNIST, Fashion-MNIST datasets, and neuromorphological N-MNIST, DVS128 Gesture datasets. The experimental results show that the accuracy of this method in static and neuromorphological datasets exceeds the current advanced methods, and the time window used is small. This work provides a new perspective for directly training high-performance SNN.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7

Similar content being viewed by others

Availability of data and materials

Not applicable.

References

  1. Tavanaei, A., Maida, A.: Bio-inspired multi-layer spiking neural network extracts discriminative features from speech signals. In: International conference on neural information processing. Springer, pp. 899–908 (2017)

  2. Stromatias, E., Neil, D., Pfeiffer, M., Galluppi, F., Furber, S.B., Liu, S.-C.: Robustness of spiking deep belief networks to noise and reduced bit precision of neuro-inspired hardware platforms. Front. Neurosci. 9, 222 (2015)

    Article  Google Scholar 

  3. Kim, S., Park, S., Na, B., Yoon, S.: Spiking-yolo: spiking neural network for energy-efficient object detection. In: Proceedings of the AAAI conference on artificial intelligence, pp. 11 270–11 277 (2020)

  4. Chen, R., Ma, H., Guo, P., Xie, S., Li, P., Wang, D.: Low latency spiking convnets with restricted output training and false spike inhibition. In: 2018 International joint conference on neural networks (IJCNN). IEEE, pp. 1–8 (2018)

  5. Hu, J., Tang, H., Tan, K.C., Li, H., Shi, L.: A spike-timing-based integrated model for pattern recognition. Neural Comput. 25(2), 450–472 (2013)

    Article  MathSciNet  MATH  Google Scholar 

  6. Hodgkin, A.L., Huxley, A.F.: A quantitative description of membrane current and its application to conduction and excitation in nerve. J. Physiol. 117(4), 500 (1952)

    Article  Google Scholar 

  7. Kasabov, N.K.: Neucube: A spiking neural network architecture for mapping, learning and understanding of spatio-temporal brain data. Neural Netw. 52, 62–76 (2014)

    Article  Google Scholar 

  8. Panchev, C., Wermter, S.: Spike-timing-dependent synaptic plasticity: from single spikes to spike trains. Neurocomputing 58, 365–371 (2004)

    Article  Google Scholar 

  9. Niu, L.-Y., Wei, Y., Long, J.-Y., Liu, W.-B.: High-accuracy spiking neural network for objective recognition based on proportional attenuating neuron. Neural Process. Lett. 54(2), 1055–1073 (2022)

    Article  Google Scholar 

  10. Amit, D.J., Brunel, N.: Dynamics of a recurrent network of spiking neurons before and following learning. Netw. Comput. Neural Syst. 8(4), 373–404 (1997)

    Article  MATH  Google Scholar 

  11. Pei, J., Deng, L., Song, S., Zhao, M., Zhang, Y., Wu, S., Wang, G., Zou, Z., Wu, Z., He, W., et al.: Towards artificial general intelligence with hybrid tianjic chip architecture. Nature 572(7767), 106–111 (2019)

    Article  Google Scholar 

  12. Roy, K., Jaiswal, A., Panda, P.: Towards spike-based machine intelligence with neuromorphic computing. Nature 575(7784), 607–617 (2019)

    Article  Google Scholar 

  13. Akopyan, F., Sawada, J., Cassidy, A., Alvarez-Icaza, R., Arthur, J., Merolla, P., Imam, N., Nakamura, Y., Datta, P., Nam, G.-J., et al.: Truenorth: Design and tool flow of a 65 mw 1 million neuron programmable neurosynaptic chip. IEEE Trans. Comput. Aided Des. Integr. Circuits Syst. 34(10), 1537–1557 (2015)

    Article  Google Scholar 

  14. O’Connor, P., Welling, M.: Deep spiking networks. arXiv preprintarXiv:1602.08323 (2016)

  15. Diehl, P.U., Neil, D., Binas, J., Cook, M., Liu, S.-C., Pfeiffer, M.: Fast-classifying, high-accuracy spiking deep networks through weight and threshold balancing. In 2015 International joint conference on neural networks (IJCNN). IEEE, p. 1–8 (2015)

  16. Sengupta, A., Ye, Y., Wang, R., Liu, C., Roy, K.: Going deeper in spiking neural networks: Vgg and residual architectures. Front. Neurosci. 13, 95 (2019)

    Article  Google Scholar 

  17. Hazan, H., Saunders, D., Sanghavi, D.T., Siegelmann, H., Kozma, R.: Unsupervised learning with self-organizing spiking neural networks. In:2018 International joint conference on neural networks (IJCNN). IEEE, pp. 1–6 (2018)

  18. Mozafari, M., Ganjtabesh, M., Nowzari-Dalini, A., Thorpe, S.J., Masquelier, T.: Bio-inspired digit recognition using reward-modulated spike-timing-dependent plasticity in deep convolutional networks. Pattern Recognit. 94, 87–95 (2019)

    Article  Google Scholar 

  19. Masquelier, T., Thorpe, S.J.: Unsupervised learning of visual features through spike timing dependent plasticity. PLoS Comput. Biol. 3(2), e31 (2007)

    Article  Google Scholar 

  20. Diehl, P.U., Cook, M.: Unsupervised learning of digit recognition using spike-timing-dependent plasticity. Front. Comput. Neurosci. 9, 99 (2015)

    Article  Google Scholar 

  21. Miao, Y., Tang, H., Pan, G.: A supervised multi-spike learning algorithm for spiking neural networks. In 2018 International joint conference on neural networks (IJCNN). IEEE, pp. 1–7 (2018)

  22. Legenstein, R., Chase, S.M., Schwartz, A.B., Maass, W.: A reward-modulated hebbian learning rule can explain experimentally observed network reorganization in a brain control task. J. Neurosci. 30(25), 8400–8410 (2010)

    Article  Google Scholar 

  23. Matsubara, T.: Spike timing-dependent conduction delay learning model classifying spatio-temporal spike patterns. In: 2017 International joint conference on neural networks (IJCNN). IEEE, pp. 1831–1839 (2017)

  24. Saunders, D.J., Siegelmann, H.T., Kozma, R.: Stdp learning of image patches with convolutional spiking neural networks. In: International joint conference on neural networks (IJCNN). IEEE pp. 1–7 (2018)

  25. Han, B., Srinivasan, G., Roy, K.: Rmp-snn: Residual membrane potential neuron for enabling deeper high-accuracy and low-latency spiking neural network. In:Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 13 558–13 567 (2020)

  26. Cao, Y., Chen, Y., Khosla, D.: Spiking deep convolutional neural networks for energy-efficient object recognition. Int. J. Comput. Vis. 113(1), 54–66 (2015)

    Article  MathSciNet  Google Scholar 

  27. Rueckauer, B., Lungu, I.-A., Hu, Y., Pfeiffer, M., Liu, S.-C.: Conversion of continuous-valued deep networks to efficient event-driven networks for image classification. Front. Neurosci. 11, 682 (2017)

    Article  Google Scholar 

  28. Tan, W., Patel, D., Kozma, R.: Strategy and benchmark for converting deep q-networks to event-driven spiking neural networks. Proc. AAAI Conf. Artif. Intell. 35(11), 9816–9824 (2021)

  29. Zhang, L., Zhou, S., Zhi, T., Du, Z., Chen, Y.: Tdsnn: from deep neural networks to deep spike neural networks with temporal-coding. Proc. AAAI Conf. Artif. Intell. 33(01), 1319–1326 (2019)

    Google Scholar 

  30. Rathi, N., Srinivasan, G., Panda, P., Roy, K.: Enabling deep spiking neural networks with hybrid conversion and spike timing dependent backpropagation. arXiv preprintarXiv:2005.01807 (2020)

  31. Bellec, G., Scherr, F., Hajek, E., Salaj, D., Legenstein, R., Maass, W.: Biologically inspired alternatives to backpropagation through time for learning in recurrent neural nets. arXiv preprintarXiv:1901.09049 (2019)

  32. Wu, Y., Deng, L., Li, G., Zhu, J., Xie, Y., Shi, L.: Direct training for spiking neural networks: faster, larger, better. Proc. AAAI Conf. Artif. Intell. 33(01), 1311–1318 (2019)

    Google Scholar 

  33. Mostafa, H.: Supervised learning based on temporal coding in spiking neural networks. IEEE Trans. Neural Netw. Learn. Syst. 29(7), 3227–3235 (2017)

    Google Scholar 

  34. Huh, D., Sejnowski, T.J.: Gradient descent for spiking neural networks’’. Adv. Neural Inf. Process. Syst. 31, 1–10 (2018)

    Google Scholar 

  35. Zheng, H., Wu, Y., Deng, L., Hu, Y., Li, G.: Going deeper with directly-trained larger spiking neural networks. In: Proceedings of the AAAI conference on artificial intelligence, pp. 11 062–11 070 (2021)

  36. Fang, W., Yu, Z., Chen, Y., Masquelier, T., Huang, T., Tian, Y.: Incorporating learnable membrane time constant to enhance learning of spiking neural networks. In:Proceedings of the IEEE/CVF international conference on computer vision, pp. 2661–2671 (2021)

  37. Zhang, D., Yan, X., She, L., Wen, Y., Poo, M.m.: Global enhancement of cortical excitability following coactivation of large neuronal populations. In: Proceedings of the national academy of sciences, pp. 20 254–20 264, (2020)

  38. Zheng, H., Wu, Y., Deng, L., Hu, Y., Li, G.: Going deeper with directly-trained larger spiking neural networks. In: Proceedings of the AAAI conference on artificial intelligence, pp. 11 062–11 070 (2021)

  39. Lu, S., Sengupta, A.: Exploring the connection between binary and spiking neural networks. Front. Neurosci. 14, 535 (2020)

    Article  Google Scholar 

  40. Hunsberger, E., Eliasmith, C.: Spiking deep networks with lif neurons. arXiv preprintarXiv:1510.08829, (2015)

  41. Haeng Lee, J., Delbruck, T., Pfeiffer, M.: Training deep spiking neural networks using backpropagation. arXiv e-prints, pp. arXiv–1608, (2016)

  42. Wu, Y., Deng, L., Li, G., Zhu, J., Shi, L.: Spatio-temporal backpropagation for training high-performance spiking neural networks. Front. Neurosci. 12, 331 (2018)

  43. Shrestha, S.B., Orchard, G.: Slayer: spike layer error reassignment in time. In: Advances in neural information processing systems

  44. Cheng, X., Hao, Y., Xu, J., Xu, B.: Lisnn: Improving spiking neural networks with lateral interactions for robust object recognition. In:IJCAI, pp. 1519–1525 (2020)

  45. Zhang, W., Li, P.: Spike-train level backpropagation for training deep recurrent spiking neural networks. In: Advances in neural information processing systems, vol. 32, (2019)

  46. Orchard, G., Jayawant, A., Cohen, G.K., Thakor, N.: Converting static image datasets to spiking neuromorphic datasets using saccades. Front. Neurosci. 9, 437 (2015)

    Article  Google Scholar 

  47. Kaiser, J., Mostafa, H., Neftci, E.: Synaptic plasticity dynamics for deep continuous local learning (decolle). Front. Neurosci. 14, 424 (2020)

    Article  Google Scholar 

  48. He, W., Wu, Y., Deng, L., Li, G., Wang, H., Tian, Y., Ding, W., Wang, W., Xie, Y.: Comparing snns and rnns on neuromorphic vision datasets: similarities and differences. Neural Netw. 132, 108–120 (2020)

  49. Bi, Y., Chadha, A., Abbas, A., Bourtsoulatze, E., Andreopoulos, Y.: Graph-based object classification for neuromorphic vision sensing. In: Proceedings of the IEEE/CVF international conference on computer vision, pp. 491–501 (2019)

  50. Sironi, A., Brambilla, M., Bourdis, N., Lagorce, X., Benosman, R.: Hats: histograms of averaged time surfaces for robust event-based object classification. In:Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1731–1740 (2018)

  51. Xing, Y., Di Caterina, G., Soraghan, J.: A new spiking convolutional recurrent neural network (scrnn) with applications to event-based hand gesture recognition. Front. Neurosci. 14, 1143 (2020)

Download references

Funding

This work is supported by National Nature Science Foundation of China (grant No. 61871106), China (grant No. 2020JH2/10100029), and the Open Project Program Foundation of the Key Laboratory of Opto-Electronics Information Processing, Chinese Academy of Sciences (OEIPO-202002).

Author information

Authors and Affiliations

Authors

Contributions

Li-Ye Niu wrote the main manuscript text and prepared figures. Ying Wei put forward some suggestions for revising the experimental design of the article. Yue Liu proofread the grammar and typesetting of the article. Jun-Yu Long and Wen-Bo Liu draw the table of the article. All authors reviewed the manuscript.

Corresponding author

Correspondence to Ying Wei.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Ethical Approval

Not applicable.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Below is the link to the electronic supplementary material.

Supplementary file 1 (pdf 269 KB)

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Niu, LY., Wei, Y., Liu, Y. et al. Spatiotemporal Backpropagation based on Channel Reward for Training High-Precision Spiking Neural Network. SIViP 17, 3467–3476 (2023). https://doi.org/10.1007/s11760-023-02569-0

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11760-023-02569-0

Keywords

Navigation