skip to main content
10.1145/3287624.3287719acmconferencesArticle/Chapter ViewAbstractPublication PagesaspdacConference Proceedingsconference-collections
research-article

Efficient FPGA implementation of local binary convolutional neural network

Published:21 January 2019Publication History

ABSTRACT

Binarized Neural Networks (BNN) has shown a capability of performing various classification tasks while taking advantage of computational simplicity and memory saving. The problem with BNN, however, is a low accuracy on large convolutional neural networks (CNN). Local Binary Convolutional Neural Network (LBCNN) compensates accuracy loss of BNN by using standard convolutional layer together with binary convolutional layer and can achieve as high accuracy as standard AlexNet CNN. For the first time we propose FPGA hardware design architecture of LBCNN and address its unique challenges. We present performance and resource usage predictor along with design space exploration framework. Our architecture on LBCNN AlexNet shows 76.6% higher performance in terms of GOPS, 2.6X and 2.7X higher performance density in terms of GOPS/Slice, and GOPS/DSP compared to previous FPGA implementation of standard AlexNet CNN.

References

  1. K. Hwang et al., "Fixed-point feedforward deep neural network design using weights," in 2014 IEEE Workshop on Signal Processing Systems (SiPS), Oct 2014, pp. 1--6.Google ScholarGoogle Scholar
  2. M. Courbariaux et al., "Binaryconnect: Training deep neural networks with binary weights during propagations," vol. 28, 11 2015.Google ScholarGoogle Scholar
  3. M. Courbariaux et al., "Binarynet: Training deep neural networks with weights and activations constrained to +1 or -1," CoRR, vol. abs/1602.02830, 2016. {Online}. Available: http://arxiv.org/abs/1602.02830Google ScholarGoogle Scholar
  4. M. Rastegari et al., "Xnor-net: Imagenet classification using binary convolutional neural networks," CoRR, vol. abs/1603.05279, 2016.Google ScholarGoogle Scholar
  5. M. Shimoda et al., "All binarized convolutional neural network and its implementation on an fpga," in 2017 International Conference on Field Programmable Technology (ICFPT), Dec 2017, pp. 291--294.Google ScholarGoogle Scholar
  6. K. Ando et al., "Brein memory: A single-chip binary/ternary reconfigurable in-memory deep neural network accelerator achieving 1.4 tops at 0.6 w," IEEE Journal of Solid-State Circuits, vol. 53, no. 4, pp. 983--994, April 2018.Google ScholarGoogle ScholarCross RefCross Ref
  7. H. Yonekawa et al., "On-chip memory based binarized convolutional deep neural network applying batch normalization free technique on an fpga," in IEEE IPDPSW, May 2017, pp. 98--105.Google ScholarGoogle Scholar
  8. S. Liang et al., "Fp-bnn: Binarized neural network on fpga," vol. 275, 10 2017.Google ScholarGoogle Scholar
  9. S. Zhou et al., "Dorefa-net: Training low bitwidth convolutional neural networks with low bitwidth gradients," CoRR, vol. abs/1606.06160, 2016. {Online}. Available: http://arxiv.org/abs/1606.06160Google ScholarGoogle Scholar
  10. M. Courbariaux et al., "Binarynet: Training deep neural networks with weights and activations constrained to +1 or -1," CoRR, vol. abs/1602.02830, 2016. {Online}. Available: http://arxiv.org/abs/1602.02830Google ScholarGoogle Scholar
  11. R. Zhao et al., "Accelerating binarized convolutional neural networks with software-programmable fpgas," ser. FPGA '17. New York, NY, USA: ACM, 2017, pp. 15--24. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. F. Juefei-Xu et al., "Local binary convolutional neural networks," in IEEE Computer Vision and Pattern Recognition (CVPR), July 2017.Google ScholarGoogle Scholar
  13. X. Lin et al., "Towards accurate binary convolutional neural network," CoRR, vol. abs/1711.11294, 2017. {Online}. Available: http://arxiv.org/abs/1711.11294 Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. F. Juefei-Xu et al., "Perturbative neural networks," in IEEE Computer Vision and Pattern Recognition (CVPR), June 2018.Google ScholarGoogle Scholar
  15. C. Zhang et al., "Optimizing fpga-based accelerator design for deep convolutional neural networks," ser. FPGA '15. New York, NY, USA: ACM, 2015, pp. 161--170. {Online}. Available Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. A. Rahman et al., "Efficient fpga acceleration of convolutional neural networks using logical-3d compute array," in 2016 Design, Automation Test in Europe Conference Exhibition (DATE), March 2016, pp. 1393--1398. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Z. Du et al., "Shidiannao: Shifting vision processing closer to the sensor," in 2015 ACM/IEEE 42nd Annual International Symposium on Computer Architecture (ISCA), June 2015, pp. 92--104. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. L. Cavigelli et al., "Origami: A 803-gop/s/w convolutional network accelerator," IEEE Transactions on Circuits and Systems for Video Technology, vol. 27, no. 11, pp. 2461--2475, Nov 2017.Google ScholarGoogle ScholarCross RefCross Ref
  19. Y. H. Chen et al., "Eyeriss: An energy-efficient reconfigurable accelerator for deep convolutional neural networks," IEEE Journal of Solid-State Circuits, vol. 52, no. 1, pp. 127--138, Jan 2017.Google ScholarGoogle ScholarCross RefCross Ref
  20. J. Qiu et al., "Going deeper with embedded fpga platform for convolutional neural network," ser. FPGA '16. New York, NY, USA: ACM, 2016, pp. 26--35. Google ScholarGoogle ScholarDigital LibraryDigital Library

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Conferences
    ASPDAC '19: Proceedings of the 24th Asia and South Pacific Design Automation Conference
    January 2019
    794 pages
    ISBN:9781450360074
    DOI:10.1145/3287624

    Copyright © 2019 ACM

    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 21 January 2019

    Permissions

    Request permissions about this article.

    Request Permissions

    Check for updates

    Qualifiers

    • research-article

    Acceptance Rates

    Overall Acceptance Rate466of1,454submissions,32%

    Upcoming Conference

    ASPDAC '25

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader