Original contribution
Weight quantization in Boltzmann machines

https://doi.org/10.1016/0893-6080(91)90077-IGet rights and content

Abstract

Hardware implementations of neural networks promise high computing power at moderate system complexity as compared to software simulations. Their design, however, must consider constraints imposed by currently available technology. Quantization of interconnection weights is the single most important such constraint. In this work, we describe simulations carried out for neural networks based on the Boltzmann Machine paradigm, on the impact of weight discretization, and choice of network architecture on performance. Our results show that this type of network is well-suited for operation with only a small number of weight levels.

References (7)

  • D.H. Ackley et al.

    Cognitive Science

    (1985)
  • J. Alspector et al.

    Advanced research in VLSI

  • N.H. Farhat

    IEEE Circ. and Dev. Magazine

    (1989, September)
There are more references available in the full text version of this article.

Cited by (34)

  • Pruning and quantization for deep neural network acceleration: A survey

    2021, Neurocomputing
    Citation Excerpt :

    Compressing CNNs by reducing precision values has been previously proposed. Converting floating-point parameters into low numerical precision datatypes for quantizing neural networks was proposed as far back as the 1990s [67,14]. Renewed interest in quantization began in the 2010s when 8-bit weight values were shown to accelerate inference without a significant drop in accuracy [233].

  • Graph Structure Learning-Based Compression Method for Convolutional Neural Networks

    2024, Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
View all citing articles on Scopus
View full text