skip to main content
10.1145/3394885.3431533acmconferencesArticle/Chapter ViewAbstractPublication PagesaspdacConference Proceedingsconference-collections
research-article

Efficient Accuracy Recovery in Approximate Neural Networks by Systematic Error Modelling

Published:29 January 2021Publication History

ABSTRACT

Approximate Computing is a promising paradigm for mitigating the computational demands of Deep Neural Networks (DNNs), by leveraging DNN performance and area, throughput or power. The DNN accuracy, affected by such approximations, can be then effectively improved through retraining. In this paper, we present a novel methodology for modelling the approximation error introduced by approximate hardware in DNNs, which accelerates retraining and achieves negligible accuracy loss. To this end, we implement the behavioral simulation of several approximate multipliers and model the error generated by such approximations on pre-trained DNNs for image classification on CIFAR10 and ImageNet. Finally, we optimize the DNN parameters by applying our error model during DNN retraining, to recover the accuracy lost due to approximations. Experimental results demonstrate the efficiency of our proposed method for accelerated retraining (11 x faster for CIFAR10 and 8x faster for ImageNet) for full DNN approximation, which allows us to deploy approximate multipliers with energy savings of up to 36% for 8-bit precision DNNs with an accuracy loss lower than 1%.

References

  1. Martín Abadi, Ashish Agarwal, Paul Barham, et al. 2015. TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems. https://www.tensorflow.org/Google ScholarGoogle Scholar
  2. Chris M. Bishop. 1995. Training with Noise is Equivalent to Tikhonov Regularization. Neural Comput. (1995).Google ScholarGoogle Scholar
  3. C. De la Parra, A. Guntoro, and A. Kumar. 2020. ProxSim: GPU-based Simulation Framework for Cross-Layer Approximate DNN Optimization. In DATE '20.Google ScholarGoogle Scholar
  4. Suyog Gupta, Ankur Agrawal, Kailash Gopalakrishnan, and Pritish Narayanan. 2015. Deep Learning with Limited Numerical Precision. ICML '15 (2015).Google ScholarGoogle Scholar
  5. I. Hammad et al. 2019. Deep Learning Training with Simulated Approximate Multipliers. In ROBIO.Google ScholarGoogle Scholar
  6. Issam Hammad and Kamal El-Sankary. 2018. Impact of Approximate Multipliers on VGG Deep Learning Network. IEEE Access (2018).Google ScholarGoogle ScholarCross RefCross Ref
  7. Song Han et al. 2016. Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding. In ICLR '16.Google ScholarGoogle Scholar
  8. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2015. Deep Residual Learning for Image Recognition. CVPR '15 (2015).Google ScholarGoogle Scholar
  9. Xin He et al. 2018. AxTrain: Hardware-Oriented Neural Network Training for Approximate Inference. ISLPED '18 (2018).Google ScholarGoogle Scholar
  10. Yihui He, Xiangyu Zhang, and Jian Sun. 2017. Channel Pruning for Accelerating Very Deep Neural Networks. ICCV '17 (2017).Google ScholarGoogle Scholar
  11. Alex Krizhevsky. 2009. Learning Multiple Layers of Features from Tiny Images. University of Toronto (2009).Google ScholarGoogle Scholar
  12. Yann LeCun and Corinna Cortes. 2010. MNIST handwritten digit database. (2010). http://yann.lecun.com/exdb/mnist/Google ScholarGoogle Scholar
  13. Darryl Dexu Lin, Sachin S. Talathi, and V. Sreekanth Annapureddy. 2016. Fixed Point Quantization of Deep Convolutional Networks. ICML 16 (2016).Google ScholarGoogle Scholar
  14. M.A.Hanif et al. 2018. Error resilience analysis for systematically employing approximate computing in convolutional neural networks. DATE 18 (2018).Google ScholarGoogle Scholar
  15. Alberto Marchisio, Muhammad Hanif, Maurizio Martina, and Muhammad Shafique. [n.d.].Google ScholarGoogle Scholar
  16. Vojtech Mrazek et al. 2016. Design of Power-efficient Approximate Multipliers for Approximate Artificial Neural Networks. In ICCAD '16.Google ScholarGoogle Scholar
  17. Vojtěch Mrázek, Radek Hrbáček, Zdeněk Vašíček, and Lukáš Sekanina. 2017. EvoApprox8B: Library of Approximate Adders and Multipliers for Circuit Design and Benchmarking of Approximation Methods. In DATE '17.Google ScholarGoogle ScholarCross RefCross Ref
  18. Vojtech Mrazek, Zdenek Vasícek, Lukás Sekanina, Muhammad Abdullah Hanif, and Muhammad Shafique. 2019. ALWANN: Automatic Layer-Wise Approximation of Deep Neural Network Accelerators without Retraining. ICCAD '19 (2019).Google ScholarGoogle ScholarCross RefCross Ref
  19. John Nickolls, Ian Buck, Michael Garland, and Kevin Skadron. 2008. Scalable Parallel Programming with CUDA. Queue (2008).Google ScholarGoogle Scholar
  20. E. Park, J. Ahn, and S. Yoo. 2017. Weighted-Entropy-Based Quantization for Deep Neural Networks. In CVPR '17.Google ScholarGoogle Scholar
  21. Olga Russakovsky et al. 2015. ImageNet Large Scale Visual Recognition Challenge. IJCV '15 (2015).Google ScholarGoogle Scholar
  22. Syed Shakib Sarwar et al. 2018. Energy-Efficient Neural Computing with Approximate Multipliers. J. Emerg. Technol. Comput. Syst. (2018).Google ScholarGoogle Scholar
  23. Jost Tobias Springenberg, Alexey Dosovitskiy, Thomas Brox, and Martin A. Riedmiller. 2014. Striving for Simplicity: The All Convolutional Net. (2014).Google ScholarGoogle Scholar
  24. Salim Ullah, Sanjeev Sripadraj Murthy, and Akash Kumar. 2018. SMApproxLib: Library of FPGA-based Approximate Multipliers. In DAC 18.Google ScholarGoogle Scholar
  25. Swagath Venkataramani et al. 2014. AxNN: Energy-efficient neuromorphic systems using approximate computing. ISLPED '14 (2014).Google ScholarGoogle Scholar
  26. Qian Zhang et al. 2015. ApproxANN: An Approximate Computing Framework for Artificial Neural Network. In DATE '15.Google ScholarGoogle Scholar

Index Terms

  1. Efficient Accuracy Recovery in Approximate Neural Networks by Systematic Error Modelling

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Conferences
          ASPDAC '21: Proceedings of the 26th Asia and South Pacific Design Automation Conference
          January 2021
          930 pages
          ISBN:9781450379991
          DOI:10.1145/3394885

          Copyright © 2021 ACM

          © 2021 Association for Computing Machinery. ACM acknowledges that this contribution was authored or co-authored by an employee, contractor or affiliate of a national government. As such, the Government retains a nonexclusive, royalty-free right to publish or reproduce this article, or to allow others to do so, for Government purposes only.

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 29 January 2021

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • research-article
          • Research
          • Refereed limited

          Acceptance Rates

          ASPDAC '21 Paper Acceptance Rate111of368submissions,30%Overall Acceptance Rate466of1,454submissions,32%

          Upcoming Conference

          ASPDAC '25

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader