Elsevier

Neurocomputing

Volume 441, 21 June 2021, Pages 52-63
Neurocomputing

Stable and compact design of Memristive GoogLeNet Neural Network

https://doi.org/10.1016/j.neucom.2021.01.122Get rights and content

Abstract

According to the requirements of edge intelligence for circuit volume, power consumption and computing performance, a Memristive GoogLeNet Neural Network (MGNN) circuit is designed using memristor which is a new device integrating storage and computing as the basic circuit element. This circuit adopts 1×1 convolution and multi-scale convolution feature fusion to reduce the number of layers required by the network while ensuring the recognition accuracy of circuit. In order to reduce the size of the memristor crossbars in the circuit, we design word-line pruning and bit-line pruning methods of Memristive Convolution (MC) layers. We also use the parameter distribution of the memristive neural network to further reduce the size of memristor crossbars. The Memristive Batch Normalization (MBN) layer and Memristive Dropout (MD) are merged into front MC layers according mathematical analysis for cutting the number of network layers and decreasing the power consumption of the circuit. We also design the channel optimization and layer optimization methods of MC layers which greatly reduce the negative effect of multi-state conductance of memristors on the accuracy, improve the stability of the circuit, and reduce the circuit volume and power consumption. Experiments show that this circuit can get 89.83% accuracy on the CIFAR-10 data set, and the power consumption of a single neuron is only 1.3μW. When the number of memristor multi-state conductance is 24=16, the accuracy of the MGNN circuit close to float MGNN can still be obtained.

Section snippets

Preface

In recent years, benefiting from technology progresses in algorithms, computing power, and data sets, deep learning has been developed tremendously in various fields such as security, e-commerce, manufacturing, agriculture, and smart home, and improved the efficiency of human production and life greatly [1]. Current intelligent applications based on deep learning usually rely on cloud data centers with powerful computing capabilities because of a lot of requirement of calculations [2], [3].

Related work

Memristor with the characteristics of variable resistance, low non-volatile power consumption and high integration density has very good application prospects in the fields of storage, artificial neural network and logic computing.

The development of neuromorphic computing circuits based on memristors can be divided into three stages: The first stage is the development of a single device. In 2008, Professor Stan William of Hewlett–Packard Lab produced a memristor in the laboratory at the first

Model of memristor

Academician Leon Chua (UC Berkeley) first proposed the concept of memristors in 1976 [13]. Since Hewlett–Packard Labs proposed the TiO2 physical realization and mathematical model of the memristor in 2008 [46], [47], the research on the memristor has become a research hot-spot in academia and industry. Research teams all over the world are trying to use various materials to prepare new devices with memristor characteristics. With the production of memristors with different materials and

Overview of MGNN

Deep learning networks such as AlexNet and Visual Geometry Group (VGG) obtain better recognition results from the perspective of increasing the depth of the network. However, the increase in the number of layers will bring about problems such as over-fitting, gradient disappearance, and gradient explosion. GoogLeNet improves the training effect from the perspective of using computing resources more efficiently and extracting more features under the same amount of calculation. Since GoogLeNet

Optimization of Batch Normalization layer

Memristive Batch Normalization layers usually follow the memristive convolution layer. MBN layers are used to speed up training process, reduce over-fitting, and make the network in-sensitive to conductance initialization [52]. MBN layers try to normalize the output of each memristive convolution layer to data with mean 0 and variance 1. The output of the kth MBN layer can be defined asx̂bn(k)=xbn(k)-E[xbn(k)]V[xbn(k)].In Eq. (13), the features xbn(k) extracted by the previous memristive

Experiment overview

Firstly, we established the MGNN model using the tensorflow framework on the traditional Graphics Processing Unit (GPU) server. Then we used the CIFAR-10 data set to train the model. In the training process, L2 regularization constraints are used for pruning the model. The trained parameters are imported into the MGNN circuit model established by MATLAB simulink, and the circuit model is used to analyze the image recognition accuracy, required memristor crossbars, power consumption and the

Conclusion

In this article, a new type of passive device with integrated storage and calculation name as memristor is used to design a compact and stable MGNN circuit, which adopts 1×1 convolution and multi-scale feature fusion in structure to reduce the number of memristive neural network layers and maintain the recognition accuracy of the circuit. In order to get a more compact circuit, this article designs word-line pruning and bit-line pruning methods for the memristive convolution layer, which

CRediT authorship contribution statement

Huanhuan Ran: Writing - original draft, Resources. Shiping Wen: Conceptualization. Kaibo Shi: Investigation. Tingwen Huang: Supervision.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Huanhuan Ran is with the Key Laboratory of Electronic Thin Films and Integrated Devices, University of Electronic Science and Technology of China, Chengdu?Sichuan, 610054, China.

References (55)

  • J. Park, S. Samarakoon, M. Bennis, M. Debbah, Wireless network intelligence at the...
  • Z. Zhou et al.

    Secure and efficient vehicle-to-grid energy trading in cyber physical systems: Integration of blockchain and edge computing

    IEEE Transactions on Systems, Man, and Cybernetics: Systems

    (2020)
  • S. Wen et al.

    Observer-based adaptive control for multiagent systems with unknown parameters under attacks

    IEEE Transactions on Neural Networks and Learning Systems

    (2021)
  • Z. Zhou et al.

    Edge intelligence: Paving the last mile of artificial intelligence with edge computing

    Proceedings of the IEEE

    (2019)
  • J. Chen et al.

    Deep learning with edge computing: A review

    Proceedings of the IEEE

    (2019)
  • K. Roy, A. Jaiswal, P. Panda, Towards spike-based machine intelligence with neuromorphic computing, Nature 575...
  • Z. Yu, A.M. Abdulghani, A. Zahid, H. Heidari, M.A. Imran, Q.H. Abbasi, An overview of neuromorphic computing for...
  • C.D. Schuman, T.E. Potok, R.M. Patton, J.D. Birdwell, M.E. Dean, G.S. Rose, J.S. Plank, A survey of neuromorphic...
  • L.O. Chua et al.

    Memristive devices and systems

    Proceedings of the IEEE

    (1976)
  • S.H. Jo, T. Chang, I. Ebong, B.B. Bhadviya, P. Mazumder, W. Lu, Nanoscale memristor device as synapse in neuromorphic...
  • S.-Y. Sun et al.

    Cascaded architecture for memristor crossbar array based larger-scale neuromorphic computing

    IEEE Access

    (2019)
  • S.-Y. Sun et al.

    Cascaded neural network for memristor based neuromorphic computing

  • S.H. Jo et al.

    Nanoscale memristor device as synapse in neuromorphic systems

    Nano Letters

    (2010)
  • P. Chi et al.

    Prime: A novel processing-in-memory architecture for neural network computation in reram-based main memory

    ACM SIGARCH Computer Architecture News

    (2016)
  • S. Wen et al.

    CKFO: Convolutional kernel first operated algorithm with applications in memristor-based convolutional neural networks

    IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems

    (2020)
  • H. Ran et al.

    Memristor-based edge computing of blaze block for image recognition

    IEEE Transactions on Neural Networks and Learning Systems

    (2020)
  • I. Kataeva et al.

    Efficient training algorithms for neural networks based on memristive crossbar circuits

  • Cited by (22)

    • Cancelable ECG biometric based on combination of deep transfer learning with DNA and amino acid approaches for human authentication

      2022, Information Sciences
      Citation Excerpt :

      These models have already learned to extract informative and powerful ECG features. There are several available pre-trained networks trained such as resnet [39], googlenet [40], xception [41] and vgg [42]. In this study, we worked on vgg-16 pre-train model [42].

    View all citing articles on Scopus

    Huanhuan Ran is with the Key Laboratory of Electronic Thin Films and Integrated Devices, University of Electronic Science and Technology of China, Chengdu?Sichuan, 610054, China.

    Shipping Wen is with Australian Artificial Intelligence Institute, Faculty of Engineering and Information Technology, University of Technology Sydney, Ultimo, NSW 2007, Australia.

    Kaibo Shi is with School of Information Science and Engineering, Chengdu University, Chengdu, Sichuan, China.

    Tingwen Huang is with Science Program, Texas A&M University at Qatar, 23874, Doha, Qatar.

    View full text