skip to main content
10.1145/3489517.3530532acmconferencesArticle/Chapter ViewAbstractPublication PagesdacConference Proceedingsconference-collections
research-article
Public Access

Towards resilient analog in-memory deep learning via data layout re-organization

Published: 23 August 2022 Publication History

Abstract

Processing in-memory paves the way for neural network inference engines. An arising challenge is to develop the software/hardware interface to automatically compile deep learning models onto in-memory computing platforms. In this paper, we observe that the data layout organization of a deep neural network (DNN) model directly impacts the model's classification accuracy. This stems from that the resistive parasitics within a crossbar introduces a dependency between the matrix data and the precision of the analog computation. To minimize the impact of the parasitics, we first perform a case study to understand the underlying matrix properties that result in computation with low and high precision, respectively. Next, we propose the XORG framework that performs data layout organization for DNNs deployed on in-memory computing platforms. The data layout organization improves precision by optimizing the weight matrix to crossbar assignments at compile time. The experimental results show that the XORG framework improves precision with up to 3.2X and 31% on the average. When accelerating DNNs using XORG, the write bit-accuracy requirements are relaxed with 1-bit and the robustness to random telegraph noise (RTN) is improved.

References

[1]
S. Choi, Y. Yang, and W. Lu. Random telegraph noise and resistance switching analysis of oxide based resistive memory. Nanoscale, 6, 11 2013.
[2]
T. H. Cormen, C. Stein, R. L. Rivest, and C. E. Leiserson. Introduction to Algorithms. McGraw-Hill Higher Education, 2001.
[3]
Z. He et al. Noise injection adaption: End-to-end ReRAM crossbar non-ideal effect adaption for neural network mapping. DAC'19, pages 57:1--57:6, 2019.
[4]
M. Hu et al. Dot-product engine for neuromorphic computing: Programming 1T1M crossbar to accelerate matrix-vector multiplication. DAC'16, pages 1--6, 2016.
[5]
M. Hu et al. Memristor-based analog computation and neural network classification with a DPE. Adv. Materials, 30, 2018.
[6]
A. James, O. Krestinskaya, and L. Chua. Neuromemristive circuits for edge computing: A review. IEEE Transactions on Neural Networks and Learning Systems, PP, 02 2019.
[7]
Y. Ji et al. Bridge the gap between neural networks and neuromorphic hardware with a neural network compiler. ASPLOS '18, page 448--460, 2018.
[8]
K. Kourtis et al. Compiling Neural Networks for a Computational Memory Accelerator. arXiv e-prints, page arXiv:2003.04293, 2020.
[9]
A. Krizhevsky. Learning multiple layers of features from tiny images. Technical report, 2009.
[10]
Y. LeCun et al. Deep learning. In Nature, pages 436--444, 2015.
[11]
B. Liu et al. Reduction and IR-drop compensations techniques for reliable neuromorphic computing. ICCAD'14, pages 63--70, 2014.
[12]
S. Liu and W. Deng. Very deep convolutional neural network based image classification using small training sample size. In 2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR), pages 730--734, Nov 2015.
[13]
C. Münch, R. Bishnoi, and M. B. Tahoori. Reliable in-memory neuromorphic computing using spintronics. ASPDAC '19, page 230--236, 2019.
[14]
L. Song et al. Pipelayer: A pipelined reram-based accelerator for deep learning. HPCA'17, pages 541--552, 2017.
[15]
N. Uysal et al. Dp-map: Towards resistive dot-product engines with improved precision. In ICCAD'20, pages 1--9. IEEE, 2020.
[16]
N. Uysal et al. Xmap: Programming memristor crossbars for analog matrix-vector multiplication: Towards high precision using representable matrices. IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems, 2021.
[17]
W. A. Wulf and S. A. McKee. Hitting the memory wall: Implications of the obvious. SIGARCH Comput. Archit. News, 23(1):20--24, 1995.
[18]
L. Xia et al. Fault-tolerant training with on-line fault detection for rram-based neural computing systems. DAC'17, pages 1--6, 2017.
[19]
B. Zhang et al. Handling stuck-at-fault defects using matrix transformation for robust inference of dnns. IEEE Trans. on Computer-Aided Design of Integrated Circuits and Systems, 1(56):1--14, 2019.
[20]
Z. Zhu et al. Mixed size crossbar based RRAM CNN accelerator with overlapped mapping. ICCAD'18, pages 69:1--69:8, 2018.

Cited By

View all
  • (2025)Optimizing hardware-software co-design based on non-ideality in memristor crossbars for in-memory computingScience China Information Sciences10.1007/s11432-024-4240-x68:2Online publication date: 14-Jan-2025
  • (2024)Marvel: Towards Efficient Federated Learning on IoT DevicesComputer Networks10.1016/j.comnet.2024.110375245(110375)Online publication date: May-2024

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
DAC '22: Proceedings of the 59th ACM/IEEE Design Automation Conference
July 2022
1462 pages
ISBN:9781450391429
DOI:10.1145/3489517
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 23 August 2022

Permissions

Request permissions for this article.

Check for updates

Qualifiers

  • Research-article

Funding Sources

Conference

DAC '22
Sponsor:
DAC '22: 59th ACM/IEEE Design Automation Conference
July 10 - 14, 2022
California, San Francisco

Acceptance Rates

Overall Acceptance Rate 1,770 of 5,499 submissions, 32%

Upcoming Conference

DAC '25
62nd ACM/IEEE Design Automation Conference
June 22 - 26, 2025
San Francisco , CA , USA

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)157
  • Downloads (Last 6 weeks)12
Reflects downloads up to 05 Mar 2025

Other Metrics

Citations

Cited By

View all
  • (2025)Optimizing hardware-software co-design based on non-ideality in memristor crossbars for in-memory computingScience China Information Sciences10.1007/s11432-024-4240-x68:2Online publication date: 14-Jan-2025
  • (2024)Marvel: Towards Efficient Federated Learning on IoT DevicesComputer Networks10.1016/j.comnet.2024.110375245(110375)Online publication date: May-2024

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Login options

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media