Abstract
Memristor and its crossbar structure have been widely studied and proven to be naturally suitable for implementing vector-matrix multiplier (VMM) operation in neural networks, making it one of the ideal underlying hardware when deploying models on edge smart devices. However, the problem of receiving much useless information is common and the non-ideal characteristics will also affect the system training accuracy and efficiency. Considering these problems, We combine the contrastive learning (CL) into in-situ training process on the memristor crossbar, improving the model feature extraction capability and robustness. Meanwhile, to make the contrastive learning integrate with the crossbar better, we proposed a multi-optimization scheme on the network loss function, model deployment method, and gradient calculation process. We also proposed some compensation strategies to address the key non-ideal characteristics we analyzed and fitted. The test results show that under the scheme proposed, the model for deployment has a high accuracy value at the beginning, reaching 83.18% in only 2 epochs, and can quickly achieve an accuracy of 3.99% increase compared to the average accuracy of the existing algorithms with the energy consumption reduced by about 8 times.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data availability
The data generated and analyzed during the current study are available from the corresponding author on reasonable request.
References
Mead C (1990) Neuromorphic electronic systems. Proc IEEE 78(10):1629–1636
Schuman CD, Kulkarni SR, Parsa M, Mitchell JP, Kay B et al (2022) Opportunities for neuromorphic computing algorithms and applications. Nature Comput Sci 2(1):10–19
Marković D, Mizrahi A, Querlioz D, Grollier J (2020) Physics for neuromorphic computing. Nature Rev Phys 2(9):499–510
Indiveri G (2021) Introducing ‘neuromorphic computing and engineering’. Neuromorphic Comput Eng 1(1):010401
Upadhyay NK, Jiang H, Wang Z, Asapu S, Xia Q, Joshua Yang J (2019) Emerging memory devices for neuromorphic computing. Adv Mater Technol 4(4):1800589
Sun B, Guo T, Zhou G, Ranjan S, Jiao Y, Wei L, Zhou YN, Wu YA (2021) Synaptic devices based neuromorphic computing applications in artificial intelligence. Materials today physics 18:100393
Yu S, Shim W, Peng X, Luo Y (2021) Rram for compute-in-memory: from inference to training. IEEE Trans Circuits Syst I Regul Pap 68(7):2753–2765
Wang C, Shi G, Qiao F, Lin R, Wu S, Hu Z (2023) Research progress in architecture and application of rram with computing-in-memory. Nanoscale advances 5(6):1559–1573
Chen X, Song T, Han Y (2021) Rram-based analog in-memory computing. In: 2021 IEEE/ACM International symposium on nanoscale architectures (NANOARCH), pp 1–6. IEEE
Li Y, Wang Z, Midya R, Xia Q, Yang JJ (2018) Review of memristor devices in neuromorphic computing: materials sciences and device challenges. J Phys D Appl Phys 51(50):503002
Hu M, Graves CE, Li C, Li Y, Ge N, Montgomery E, Davila N, Jiang H, Williams RS, Yang JJ et al (2018) Memristor-based analog computation and neural network classification with a dot product engine. Adv Mater 30(9):1705914
Ankit A, Hajj IE, Chalamalasetti SR, Ndu G, Foltin M, Williams RS, Faraboschi P, Hwu W-mW, Strachan JP, Roy K et al (2019) Puma: a programmable ultra-efficient memristor-based accelerator for machine learning inference. In: Proceedings of the twenty-fourth international conference on architectural support for programming languages and operating systems, pp 715–731
Sangwan VK, Hersam MC (2020) Neuromorphic nanoelectronic materials. Nat Nanotechnol 15(7):517–528
Xia Q, Yang JJ (2019) Memristive crossbar arrays for brain-inspired computing. Nat Mater 18(4):309–323
Zidan MA, Strachan JP, Lu WD (2018) The future of electronics based on memristive systems. Nature electron 1(1):22–29
Wang Y, Wu S, Tian L, Shi L (2020) Ssm: a high-performance scheme for in situ training of imprecise memristor neural networks. Neurocomputing 407:270–280
Dickson JA, McLeod RD, Card HC (1993) Stochastic arithmetic implementations of neural networks with in situ learning. In: IEEE International conference on neural networks, pp 711–716. IEEE
Kaiser J, Borders WA, Camsari KY, Fukami S, Ohno H, Datta S (2022) Hardware-aware in situ learning based on stochastic magnetic tunnel junctions. Phys Rev Appl 17(1):014016
Yan FY, Ayers H, Zhu C, Fouladi S, Hong J, Zhang K, Levis P, Winstein K (2020) Learning in situ: a randomized experiment in video streaming. In: 17th USENIX Symposium on Networked Systems Design and Implementation (NSDI 20), pp 495–511
Zhang Q, Wu H, Yao P, Zhang W, Gao B, Deng N, Qian H (2018) Sign backpropagation: an on-chip learning algorithm for analog rram neuromorphic computing systems. Neural Netw 108:217–223
Li J, Zhou G, Li Y, Chen J, Ge Y, Mo Y, Yang Y, Qian X, Jiang W, Liu H et al (2022) Reduction 93.7% time and power consumption using a memristor-based imprecise gradient update algorithm. Artif Intell Rev 55(1):657–677
Yadav DN, Thangkhiew PL, Datta K, Chakraborty S, Drechsler R, Sengupta I (2022) Feed-forward learning algorithm for resistive memories. J Syst Architect 131:102730
Cheng Y, Wang D, Zhou P, Zhang T (2018) Model compression and acceleration for deep neural networks: the principles, progress, and challenges. IEEE Signal Process Mag 35(1):126–136
Pandiyan V, Cui D, Le-Quang T, Deshpande P, Wasmer K, Shevchik S (2022) In situ quality monitoring in direct energy deposition process using co-axial process zone imaging and deep contrastive learning. J Manuf Process 81:1064–1075
Abed-Esfahani P, Darwin BC, Howard D, Wang N, Kim E, Lerch J, French L (2022) Evaluation of deep convolutional neural networks for in situ hybridization gene expression image representation. PLoS ONE 17(1):0262717
Wang K (2023) Contrastive learning-based semantic segmentation for in-situ stratified defect detection in additive manufacturing. J Manuf Syst 68:465–476
Jiang X, Cai W, Ding Y, Wang X, Hong D, Yang Z, Gao W (2023) Camouflaged object segmentation based on joint salient object for contrastive learning. IEEE Transactions on instrumentation and measurement
Zheng L, Liu Z, Yang Y, Shen H-B (2022) Accurate inference of gene regulatory interactions from spatial gene expression with deep contrastive learning. Bioinformatics 38(3):746–753
Yang J, Kang Z, Yang Z, Xie J, Xue B, Yang J, Tao J (2022) A laboratory open-set martian rock classification method based on spectral signatures. IEEE Trans Geosci Remote Sens 60:1–15
Becker S, Hinton GE (1992) Self-organizing neural network that discovers surfaces in random-dot stereograms. Nature 355(6356):161–163
Le-Khac PH, Healy G, Smeaton AF (2020) Contrastive representation learning: a framework and review. Ieee Access 8:193907–193934
Chen T, Kornblith S, Norouzi M, Hinton G (2020) A simple framework for contrastive learning of visual representations. In: International Conference on Machine Learning, pp 1597–1607. PMLR
Jaiswal A, Babu AR, Zadeh MZ, Banerjee D, Makedon F (2020) A survey on contrastive self-supervised learning. Technologies 9(1):2
Liu X, Zhang F, Hou Z, Mian L, Wang Z, Zhang J, Tang J (2021) Self-supervised learning: generative or contrastive. IEEE Trans Knowl Data Eng 35(1):857–876
Chua L (1971) Memristor-the missing circuit element. IEEE Transactions on circuit theory 18(5):507–519
Strukov DB, Snider GS, Stewart DR, Williams RS (2008) The missing memristor found. nature 453(7191):80–83
Magee JC, Grienberger C (2020) Synaptic plasticity forms and functions. Annu Rev Neurosci 43(1):95–117
Wang Z, Wu H, Burr GW, Hwang CS, Wang KL, Xia Q, Yang JJ (2020) Resistive switching materials for information processing. Nat Rev Mater 5(3):173–195
Prezioso M, Merrikh-Bayat F, Hoskins BD, Adam GC, Likharev KK, Strukov DB (2015) Training and operation of an integrated neuromorphic network based on metal-oxide memristors. Nature 521(7550):61–64
Xi Y, Gao B, Tang J, Chen A, Chang M-F, Hu XS, Van Der Spiegel J, Qian H, Wu H (2020) In-memory learning with analog resistive switching memory: a review and perspective. Proc IEEE 109(1):14–42
Zhou G, Sun B, Hu X, Sun L, Zou Z, Xiao B, Qiu W, Wu B, Li J, Han J et al (2021) Negative photoconductance effect: an extension function of the tiox-based memristor. Adv Sci 8(13):2003765
Cortes C (1995) Support-vector networks. Machine learning
Choi S, Sheridan P, Lu WD (2015) Data clustering using memristor networks. Sci Rep 5(1):10492
Sheridan PM, Cai F, Du C, Ma W, Zhang Z, Lu WD (2017) Sparse coding with memristor networks. Nat Nanotechnol 12(8):784–789
Adhikari SP, Yang C, Slot K, Strzelecki M, Kim H (2018) Hybrid no-propagation learning for multilayer neural networks. Neurocomputing 321:28–35
Yadav DN, Datta K, Sengupta I (2020) Analyzing fault tolerance behaviour in memristor-based crossbar for neuromorphic applications. In: 2020 IEEE International test conference India, pp 1–9. IEEE
Ansari M, Fayyazi A, Kamal M, Afzali-Kusha A, Pedram M (2019) Octan: an on-chip training algorithm for memristive neuromorphic circuits. IEEE Trans Circuits Syst I Regul Pap 66(12):4687–4698
Krestinskaya O, Salama KN, James AP (2018) Learning in memristive neural network architectures using analog backpropagation circuits. IEEE Trans Circuits Syst I Regul Pap 66(2):719–732
Merkel C, Kudithipudi D (2014) A stochastic learning algorithm for neuromemristive systems. In: 2014 27th IEEE international System-on-Chip Conference (SOCC), pp 359–364. IEEE
Acknowledgements
The work was supported by National Natural Science Foundation of China (Grant No. 62476230 ), Natural Science Foundation of Chongqing(Grant No. CSTB2023NSCQ-MSX0018), Fundamental Research Funds for the Central Universities (Grant No. SWU-KR22046),National Natural Science Foundation of China (Grant No. 61976246).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflicts of interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Xiong, F., Zhou, Y., Hu, X. et al. Multi-optimization scheme for in-situ training of memristor neural network based on contrastive learning. Appl Intell 55, 99 (2025). https://doi.org/10.1007/s10489-024-05957-2
Accepted:
Published:
DOI: https://doi.org/10.1007/s10489-024-05957-2