Skip to main content

SparseMAX: Accelerating Quantum Neural Networks on GPU Clusters Using Sparse-Matrix Kernels

  • Conference paper
  • First Online:
Embedded Computer Systems: Architectures, Modeling, and Simulation (SAMOS 2021)

Abstract

The growing popularity of Applied Quantum Mechanics and Artificial Intelligence drives the need for integrating the two fields. Quantum Neural Networks (QNNs) incorporate quantum aspects into classical deep learning networks which are capable of performing universal quantum computations. The dense representation of QNNs presents great challenges in terms of computational cost and noise susceptibility. In this paper, we present SparseMAX, a novel Sparse Quantum Neural Network (SQNN) that is robust to noise and interference for large volumes of network parameters. We also introduce Quantron (\(\psi \)), a generalized version of perceptron, which acts on qubits and performs the necessary quantum operations. Based on these insights, we develop 2 GPU kernels. The first kernel estimates the network architecture through a quantum training algorithm. The second kernel accelerates a sparsified version of the network matrices on a GPU cluster. We validate our kernel performance and training algorithm and present the results in terms of inference time, GPU efficiency and scalability. On an average, SparseMAX utilizes 54.83% of our GPU cluster’s compute resources, while offering a 41.51\(\times \) speedup in terms of serial inference timing measurements for network layer range [120, 1920] and neurons per layer range [1024,4096]

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 69.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 89.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Hinton, G.E., Osindero, S., Teh, Y.: A fast learning algorithm for deep belief nets. Neural Comput. 18, 1527–1554 (2006)

    Google Scholar 

  2. Hu, H., Liu, Z., An, J.: Mining mobile intelligence for wireless systems: a deep neural network approach. IEEE Comput. Intell. Mag. 15(1), 24–31 (2020)

    Article  Google Scholar 

  3. Jeremy, K., Alford, S., Gadepally, V.: Sparse deep neural network graph challenge (2019)

    Google Scholar 

  4. Kak, S.: On quantum neural computing. Adv. Imaging Electron Phys. 94, 259–313 (1995)

    Article  Google Scholar 

  5. Chrisley, R.: Quantum learning. New directions in cognitive science. In: Proceedings of the International Symposium, Saariselka, August 1995, Lapland, Finland. Helsinki: Finnish Association of Artificial Intelligence, pp. 77–89 (1995)

    Google Scholar 

  6. Behrman, E.C., Steck, J.E., Kumar, P., Walsh, K.A.: Quantum algorithm design using dynamic learning. Quantum Inf. Comput. 8(1), 12–29 (2008)

    MathSciNet  MATH  Google Scholar 

  7. Liu, Z., Duan, L.M., Deng, D.L.: Solving quantum master equations with deep quantum neural networks. Phys. Rev. Res. 4(1), 013097 (2020)

    Google Scholar 

  8. Broughton, M., Verdon, G., McCourt, T., et al.: Tensorflow quantum: a software framework for quantum machine learning (2020)

    Google Scholar 

  9. Han, S., Pool, J., Tran, J., Dally, W.: Learning both weights and connections for efficient neural network. In: Advances in Neural Information Processing Systems, pp. 1135–1143 (2015)

    Google Scholar 

  10. Han, S., et al.: EIE: efficient inference engine on compressed deep neural network. In: Proceedings of the 43rd International Symposium on Computer Architecture, pp. 243–254 (2016)

    Google Scholar 

  11. Paula Neto, F.M.D., Ludermir, T.B., Oliveira, W.R.D., Silva, A.J.D.: Quantum perceptron with dynamic internal memory. In: International Joint Conference on Neural Networks (IJCNN), pp. 1–8 (2018)

    Google Scholar 

  12. DiAdamo, S., Notzel, J., Zanger, B., Beşe, M.M.: Qunetsim: a software framework for quantum network. IEEE Trans Quantum Eng. 2, 1–12 (2020)

    Google Scholar 

  13. Liu, W., Gao, P., Wang, Y., Yu, W., Zhang, M.: A unitary weights based one-iteration quantum perceptron algorithm for non-ideal training sets. IEEE Access 7, 36854–36865 (2019)

    Article  Google Scholar 

  14. van den Berg, E.: Iterative quantum phase estimation with optimized sample complexity. In: 2020 IEEE International Conference on Quantum Computing and Engineering (QCE), pp. 1–10 (2020)

    Google Scholar 

  15. Mohammadbagherpoor, H., Oh, Y., Dreher, P., Singh, A., Yu, X., Rindos, A.J.: An improved implementation approach for quantum phase estimation on quantum computers. In: IEEE International Conference on Rebooting Computing (ICRC), pp. 1–9 (2019)

    Google Scholar 

  16. Gardner, E.: Mathematical and general. J. Phys. 21, 257 (1988)

    MathSciNet  Google Scholar 

  17. Zhao, J.-Y.: Implementing associative memory with quantum neural networks. In: Proceedings of 2004 International Conference on Machine Learning and Cybernetics, pp. 3197–3200 (2004)

    Google Scholar 

  18. Guo, C.: Accelerating sparse DNN models without hardware-support via tile-wise sparsity. In: International Conference for High Performance Computing, Networking, Storage and Analysis, vol. 20, pp. 1–15 (2020)

    Google Scholar 

  19. Argueta, A., Chiang, D.: Accelerating sparse matrix operations in neural networks on graphics processing units. In: Accelerating Sparse Matrix Operations in Neural Networks on Graphics Processing Units, vol. 6224, no. 10, pp. 1619-1626 (2019)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Anand Ravishankar .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Ravishankar, A., Natarajan, S., Bharathi Malakreddy, A. (2022). SparseMAX: Accelerating Quantum Neural Networks on GPU Clusters Using Sparse-Matrix Kernels. In: Orailoglu, A., Jung, M., Reichenbach, M. (eds) Embedded Computer Systems: Architectures, Modeling, and Simulation. SAMOS 2021. Lecture Notes in Computer Science, vol 13227. Springer, Cham. https://doi.org/10.1007/978-3-031-04580-6_28

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-04580-6_28

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-04579-0

  • Online ISBN: 978-3-031-04580-6

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics