skip to main content
10.1145/3489517.3530619acmconferencesArticle/Chapter ViewAbstractPublication PagesdacConference Proceedingsconference-collections
research-article

Hardware-efficient stochastic rounding unit design for DNN training: late breaking results

Published: 23 August 2022 Publication History

Abstract

Stochastic rounding is crucial in the training of low-bit deep neural networks (DNNs) to achieve high accuracy. Unfortunately, prior studies require a large number of high-precision stochastic rounding units (SRUs) to guarantee the low-bit DNN accuracy, which involves considerable hardware overhead. In this paper, we propose an automated framework to explore hardware-efficient low-bit SRUs (ESRUs) that can still generate high-quality random numbers to guarantee the accuracy of low-bit DNN training. Experimental results using state-of-the-art DNN models demonstrate that, compared to the prior 24-bit SRU with 24-bit pseudo random number generator (PRNG), our 8-bit with 3-bit PRNG reduces the SRU resource usage by 9.75× while achieving a higher accuracy.

References

[1]
Courbariaux et al. 2015. Binaryconnect: Training deep neural networks with binary weights during propagations. In (NeurIPS).
[2]
Cheng Luo et al. 2020. Towards efficient deep neural network training by FPGA-based batch-level parallelism. Journal of Semiconductors (2020).
[3]
Fengfu Li et al. 2016. Ternary weight networks. arXiv preprint arXiv:1605.04711 (2016).
[4]
Roth Jr et al. 2016. Digital systems design using VHDL. Cengage Learning.
[5]
Sung-En Chang et al. 2021. Mix and match: A novel fpga-centric deep neural network quantization framework. In 2021 HPCA. IEEE.
[6]
Sung-En Chang et al. 2021. RMSMP: A Novel Deep Neural Network Quantization Framework with Row-wise Mixed Schemes and Multiple Precisions. In Proceedings of ICCV. 5251--5260.
[7]
Naigang et al. Wang. 2018. Training deep neural networks with 8-bit floating point numbers. In Proceedings of the NIPS. 7686--7695.
[8]
Yukuan et al. Yang. 2020. Training high-performance and large-scale deep neural networks with full 8-bit integers. Neural Networks (2020).
[9]
Kang et al. Zhao. 2021. Distribution Adaptive INT8 Quantization for Training CNNs. In Proceedings of the AAAI.
[10]
Shuchang et al. Zhou. 2016. Dorefa-net: Training low bitwidth Convolutional Neural Networks with low bitwidth gradients. arXiv:1606.06160 (2016).
[11]
Feng et al. Zhu. 2020. Towards unified int8 training for convolutional neural network. In Proceedings of CVPR. 1969--1979.

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
DAC '22: Proceedings of the 59th ACM/IEEE Design Automation Conference
July 2022
1462 pages
ISBN:9781450391429
DOI:10.1145/3489517
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 23 August 2022

Permissions

Request permissions for this article.

Check for updates

Qualifiers

  • Research-article

Conference

DAC '22
Sponsor:
DAC '22: 59th ACM/IEEE Design Automation Conference
July 10 - 14, 2022
California, San Francisco

Acceptance Rates

Overall Acceptance Rate 1,770 of 5,499 submissions, 32%

Upcoming Conference

DAC '25
62nd ACM/IEEE Design Automation Conference
June 22 - 26, 2025
San Francisco , CA , USA

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 243
    Total Downloads
  • Downloads (Last 12 months)44
  • Downloads (Last 6 weeks)4
Reflects downloads up to 28 Feb 2025

Other Metrics

Citations

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media