Skip to main content

Advertisement

Multi-optimization scheme for in-situ training of memristor neural network based on contrastive learning

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

Memristor and its crossbar structure have been widely studied and proven to be naturally suitable for implementing vector-matrix multiplier (VMM) operation in neural networks, making it one of the ideal underlying hardware when deploying models on edge smart devices. However, the problem of receiving much useless information is common and the non-ideal characteristics will also affect the system training accuracy and efficiency. Considering these problems, We combine the contrastive learning (CL) into in-situ training process on the memristor crossbar, improving the model feature extraction capability and robustness. Meanwhile, to make the contrastive learning integrate with the crossbar better, we proposed a multi-optimization scheme on the network loss function, model deployment method, and gradient calculation process. We also proposed some compensation strategies to address the key non-ideal characteristics we analyzed and fitted. The test results show that under the scheme proposed, the model for deployment has a high accuracy value at the beginning, reaching 83.18% in only 2 epochs, and can quickly achieve an accuracy of 3.99% increase compared to the average accuracy of the existing algorithms with the energy consumption reduced by about 8 times.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15
Fig. 16

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

Data availability

The data generated and analyzed during the current study are available from the corresponding author on reasonable request.

References

  1. Mead C (1990) Neuromorphic electronic systems. Proc IEEE 78(10):1629–1636

    Article  MATH  Google Scholar 

  2. Schuman CD, Kulkarni SR, Parsa M, Mitchell JP, Kay B et al (2022) Opportunities for neuromorphic computing algorithms and applications. Nature Comput Sci 2(1):10–19

    Article  MATH  Google Scholar 

  3. Marković D, Mizrahi A, Querlioz D, Grollier J (2020) Physics for neuromorphic computing. Nature Rev Phys 2(9):499–510

    Article  Google Scholar 

  4. Indiveri G (2021) Introducing ‘neuromorphic computing and engineering’. Neuromorphic Comput Eng 1(1):010401

    Article  MATH  Google Scholar 

  5. Upadhyay NK, Jiang H, Wang Z, Asapu S, Xia Q, Joshua Yang J (2019) Emerging memory devices for neuromorphic computing. Adv Mater Technol 4(4):1800589

    Article  Google Scholar 

  6. Sun B, Guo T, Zhou G, Ranjan S, Jiao Y, Wei L, Zhou YN, Wu YA (2021) Synaptic devices based neuromorphic computing applications in artificial intelligence. Materials today physics 18:100393

    Article  Google Scholar 

  7. Yu S, Shim W, Peng X, Luo Y (2021) Rram for compute-in-memory: from inference to training. IEEE Trans Circuits Syst I Regul Pap 68(7):2753–2765

    Article  MATH  Google Scholar 

  8. Wang C, Shi G, Qiao F, Lin R, Wu S, Hu Z (2023) Research progress in architecture and application of rram with computing-in-memory. Nanoscale advances 5(6):1559–1573

    Article  MATH  Google Scholar 

  9. Chen X, Song T, Han Y (2021) Rram-based analog in-memory computing. In: 2021 IEEE/ACM International symposium on nanoscale architectures (NANOARCH), pp 1–6. IEEE

  10. Li Y, Wang Z, Midya R, Xia Q, Yang JJ (2018) Review of memristor devices in neuromorphic computing: materials sciences and device challenges. J Phys D Appl Phys 51(50):503002

    Article  Google Scholar 

  11. Hu M, Graves CE, Li C, Li Y, Ge N, Montgomery E, Davila N, Jiang H, Williams RS, Yang JJ et al (2018) Memristor-based analog computation and neural network classification with a dot product engine. Adv Mater 30(9):1705914

    Article  Google Scholar 

  12. Ankit A, Hajj IE, Chalamalasetti SR, Ndu G, Foltin M, Williams RS, Faraboschi P, Hwu W-mW, Strachan JP, Roy K et al (2019) Puma: a programmable ultra-efficient memristor-based accelerator for machine learning inference. In: Proceedings of the twenty-fourth international conference on architectural support for programming languages and operating systems, pp 715–731

  13. Sangwan VK, Hersam MC (2020) Neuromorphic nanoelectronic materials. Nat Nanotechnol 15(7):517–528

    Article  MATH  Google Scholar 

  14. Xia Q, Yang JJ (2019) Memristive crossbar arrays for brain-inspired computing. Nat Mater 18(4):309–323

    Article  MATH  Google Scholar 

  15. Zidan MA, Strachan JP, Lu WD (2018) The future of electronics based on memristive systems. Nature electron 1(1):22–29

    Article  MATH  Google Scholar 

  16. Wang Y, Wu S, Tian L, Shi L (2020) Ssm: a high-performance scheme for in situ training of imprecise memristor neural networks. Neurocomputing 407:270–280

    Article  MATH  Google Scholar 

  17. Dickson JA, McLeod RD, Card HC (1993) Stochastic arithmetic implementations of neural networks with in situ learning. In: IEEE International conference on neural networks, pp 711–716. IEEE

  18. Kaiser J, Borders WA, Camsari KY, Fukami S, Ohno H, Datta S (2022) Hardware-aware in situ learning based on stochastic magnetic tunnel junctions. Phys Rev Appl 17(1):014016

    Article  Google Scholar 

  19. Yan FY, Ayers H, Zhu C, Fouladi S, Hong J, Zhang K, Levis P, Winstein K (2020) Learning in situ: a randomized experiment in video streaming. In: 17th USENIX Symposium on Networked Systems Design and Implementation (NSDI 20), pp 495–511

  20. Zhang Q, Wu H, Yao P, Zhang W, Gao B, Deng N, Qian H (2018) Sign backpropagation: an on-chip learning algorithm for analog rram neuromorphic computing systems. Neural Netw 108:217–223

    Article  MATH  Google Scholar 

  21. Li J, Zhou G, Li Y, Chen J, Ge Y, Mo Y, Yang Y, Qian X, Jiang W, Liu H et al (2022) Reduction 93.7% time and power consumption using a memristor-based imprecise gradient update algorithm. Artif Intell Rev 55(1):657–677

  22. Yadav DN, Thangkhiew PL, Datta K, Chakraborty S, Drechsler R, Sengupta I (2022) Feed-forward learning algorithm for resistive memories. J Syst Architect 131:102730

    Article  MATH  Google Scholar 

  23. Cheng Y, Wang D, Zhou P, Zhang T (2018) Model compression and acceleration for deep neural networks: the principles, progress, and challenges. IEEE Signal Process Mag 35(1):126–136

    Article  MATH  Google Scholar 

  24. Pandiyan V, Cui D, Le-Quang T, Deshpande P, Wasmer K, Shevchik S (2022) In situ quality monitoring in direct energy deposition process using co-axial process zone imaging and deep contrastive learning. J Manuf Process 81:1064–1075

    Article  Google Scholar 

  25. Abed-Esfahani P, Darwin BC, Howard D, Wang N, Kim E, Lerch J, French L (2022) Evaluation of deep convolutional neural networks for in situ hybridization gene expression image representation. PLoS ONE 17(1):0262717

    Article  Google Scholar 

  26. Wang K (2023) Contrastive learning-based semantic segmentation for in-situ stratified defect detection in additive manufacturing. J Manuf Syst 68:465–476

    Article  MATH  Google Scholar 

  27. Jiang X, Cai W, Ding Y, Wang X, Hong D, Yang Z, Gao W (2023) Camouflaged object segmentation based on joint salient object for contrastive learning. IEEE Transactions on instrumentation and measurement

  28. Zheng L, Liu Z, Yang Y, Shen H-B (2022) Accurate inference of gene regulatory interactions from spatial gene expression with deep contrastive learning. Bioinformatics 38(3):746–753

    Article  MATH  Google Scholar 

  29. Yang J, Kang Z, Yang Z, Xie J, Xue B, Yang J, Tao J (2022) A laboratory open-set martian rock classification method based on spectral signatures. IEEE Trans Geosci Remote Sens 60:1–15

    MATH  Google Scholar 

  30. Becker S, Hinton GE (1992) Self-organizing neural network that discovers surfaces in random-dot stereograms. Nature 355(6356):161–163

    Article  MATH  Google Scholar 

  31. Le-Khac PH, Healy G, Smeaton AF (2020) Contrastive representation learning: a framework and review. Ieee Access 8:193907–193934

    Article  Google Scholar 

  32. Chen T, Kornblith S, Norouzi M, Hinton G (2020) A simple framework for contrastive learning of visual representations. In: International Conference on Machine Learning, pp 1597–1607. PMLR

  33. Jaiswal A, Babu AR, Zadeh MZ, Banerjee D, Makedon F (2020) A survey on contrastive self-supervised learning. Technologies 9(1):2

    Article  MATH  Google Scholar 

  34. Liu X, Zhang F, Hou Z, Mian L, Wang Z, Zhang J, Tang J (2021) Self-supervised learning: generative or contrastive. IEEE Trans Knowl Data Eng 35(1):857–876

    MATH  Google Scholar 

  35. Chua L (1971) Memristor-the missing circuit element. IEEE Transactions on circuit theory 18(5):507–519

    Article  MATH  Google Scholar 

  36. Strukov DB, Snider GS, Stewart DR, Williams RS (2008) The missing memristor found. nature 453(7191):80–83

    Google Scholar 

  37. Magee JC, Grienberger C (2020) Synaptic plasticity forms and functions. Annu Rev Neurosci 43(1):95–117

    Article  MATH  Google Scholar 

  38. Wang Z, Wu H, Burr GW, Hwang CS, Wang KL, Xia Q, Yang JJ (2020) Resistive switching materials for information processing. Nat Rev Mater 5(3):173–195

    Article  MATH  Google Scholar 

  39. Prezioso M, Merrikh-Bayat F, Hoskins BD, Adam GC, Likharev KK, Strukov DB (2015) Training and operation of an integrated neuromorphic network based on metal-oxide memristors. Nature 521(7550):61–64

  40. Xi Y, Gao B, Tang J, Chen A, Chang M-F, Hu XS, Van Der Spiegel J, Qian H, Wu H (2020) In-memory learning with analog resistive switching memory: a review and perspective. Proc IEEE 109(1):14–42

    Article  MATH  Google Scholar 

  41. Zhou G, Sun B, Hu X, Sun L, Zou Z, Xiao B, Qiu W, Wu B, Li J, Han J et al (2021) Negative photoconductance effect: an extension function of the tiox-based memristor. Adv Sci 8(13):2003765

    Article  Google Scholar 

  42. Cortes C (1995) Support-vector networks. Machine learning

  43. Choi S, Sheridan P, Lu WD (2015) Data clustering using memristor networks. Sci Rep 5(1):10492

    Article  MATH  Google Scholar 

  44. Sheridan PM, Cai F, Du C, Ma W, Zhang Z, Lu WD (2017) Sparse coding with memristor networks. Nat Nanotechnol 12(8):784–789

    Article  MATH  Google Scholar 

  45. Adhikari SP, Yang C, Slot K, Strzelecki M, Kim H (2018) Hybrid no-propagation learning for multilayer neural networks. Neurocomputing 321:28–35

    Article  Google Scholar 

  46. Yadav DN, Datta K, Sengupta I (2020) Analyzing fault tolerance behaviour in memristor-based crossbar for neuromorphic applications. In: 2020 IEEE International test conference India, pp 1–9. IEEE

  47. Ansari M, Fayyazi A, Kamal M, Afzali-Kusha A, Pedram M (2019) Octan: an on-chip training algorithm for memristive neuromorphic circuits. IEEE Trans Circuits Syst I Regul Pap 66(12):4687–4698

    Article  Google Scholar 

  48. Krestinskaya O, Salama KN, James AP (2018) Learning in memristive neural network architectures using analog backpropagation circuits. IEEE Trans Circuits Syst I Regul Pap 66(2):719–732

    Article  MATH  Google Scholar 

  49. Merkel C, Kudithipudi D (2014) A stochastic learning algorithm for neuromemristive systems. In: 2014 27th IEEE international System-on-Chip Conference (SOCC), pp 359–364. IEEE

Download references

Acknowledgements

The work was supported by National Natural Science Foundation of China (Grant No. 62476230 ), Natural Science Foundation of Chongqing(Grant No. CSTB2023NSCQ-MSX0018), Fundamental Research Funds for the Central Universities (Grant No. SWU-KR22046),National Natural Science Foundation of China (Grant No. 61976246).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Xiaofang Hu.

Ethics declarations

Conflicts of interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Xiong, F., Zhou, Y., Hu, X. et al. Multi-optimization scheme for in-situ training of memristor neural network based on contrastive learning. Appl Intell 55, 99 (2025). https://doi.org/10.1007/s10489-024-05957-2

Download citation

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s10489-024-05957-2

Keywords