Loading [a11y]/accessibility-menu.js
Joint CPU-FPGA Hardware-Aware Quantized Training of Graph Convolutional Networks | IEEE Conference Publication | IEEE Xplore

Joint CPU-FPGA Hardware-Aware Quantized Training of Graph Convolutional Networks


Abstract:

In this work, a method for training Graph Convolutional Networks where the bulk of the computations for both the forward and backward passes are performed with short word...Show More

Abstract:

In this work, a method for training Graph Convolutional Networks where the bulk of the computations for both the forward and backward passes are performed with short word-length fixed-point operands on a unified FPGA hardware accelerator is presented. The accelerator targets the programmable logic of an AMD Zynq Utrascale+ FPGA device with a scalable architecture that can be configured with a variable number of hardware threads and compute units per thread. The gradients and activations are computed using a streaming architecture to reduce memory accesses and pipeline stalls while quantization is applied to all adjacency, feature, weight, and gradient tensors. Experiments show that using the hardware-aware quantized training methodology for this application, we can achieve classification accuracy with little to no classification accuracy degradation when quantizing down to 4 -bit fixed-point compared to the reference 32 -bit floating-point model. Additionally, we see a significant, up to 34x speedup, compared to running the same model on the CPU in the processing system (PS) only. The proposed hardware-software combination enables efficient, accurate, and fast training and inference at the edge.
Date of Conference: 29-30 October 2024
Date Added to IEEE Xplore: 18 November 2024
ISBN Information:
Conference Location: Lund, Sweden

Contact IEEE to Subscribe

References

References is not available for this document.