Abstract
The growing popularity of Applied Quantum Mechanics and Artificial Intelligence drives the need for integrating the two fields. Quantum Neural Networks (QNNs) incorporate quantum aspects into classical deep learning networks which are capable of performing universal quantum computations. The dense representation of QNNs presents great challenges in terms of computational cost and noise susceptibility. In this paper, we present SparseMAX, a novel Sparse Quantum Neural Network (SQNN) that is robust to noise and interference for large volumes of network parameters. We also introduce Quantron (\(\psi \)), a generalized version of perceptron, which acts on qubits and performs the necessary quantum operations. Based on these insights, we develop 2 GPU kernels. The first kernel estimates the network architecture through a quantum training algorithm. The second kernel accelerates a sparsified version of the network matrices on a GPU cluster. We validate our kernel performance and training algorithm and present the results in terms of inference time, GPU efficiency and scalability. On an average, SparseMAX utilizes 54.83% of our GPU cluster’s compute resources, while offering a 41.51\(\times \) speedup in terms of serial inference timing measurements for network layer range [120, 1920] and neurons per layer range [1024,4096]
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Hinton, G.E., Osindero, S., Teh, Y.: A fast learning algorithm for deep belief nets. Neural Comput. 18, 1527–1554 (2006)
Hu, H., Liu, Z., An, J.: Mining mobile intelligence for wireless systems: a deep neural network approach. IEEE Comput. Intell. Mag. 15(1), 24–31 (2020)
Jeremy, K., Alford, S., Gadepally, V.: Sparse deep neural network graph challenge (2019)
Kak, S.: On quantum neural computing. Adv. Imaging Electron Phys. 94, 259–313 (1995)
Chrisley, R.: Quantum learning. New directions in cognitive science. In: Proceedings of the International Symposium, Saariselka, August 1995, Lapland, Finland. Helsinki: Finnish Association of Artificial Intelligence, pp. 77–89 (1995)
Behrman, E.C., Steck, J.E., Kumar, P., Walsh, K.A.: Quantum algorithm design using dynamic learning. Quantum Inf. Comput. 8(1), 12–29 (2008)
Liu, Z., Duan, L.M., Deng, D.L.: Solving quantum master equations with deep quantum neural networks. Phys. Rev. Res. 4(1), 013097 (2020)
Broughton, M., Verdon, G., McCourt, T., et al.: Tensorflow quantum: a software framework for quantum machine learning (2020)
Han, S., Pool, J., Tran, J., Dally, W.: Learning both weights and connections for efficient neural network. In: Advances in Neural Information Processing Systems, pp. 1135–1143 (2015)
Han, S., et al.: EIE: efficient inference engine on compressed deep neural network. In: Proceedings of the 43rd International Symposium on Computer Architecture, pp. 243–254 (2016)
Paula Neto, F.M.D., Ludermir, T.B., Oliveira, W.R.D., Silva, A.J.D.: Quantum perceptron with dynamic internal memory. In: International Joint Conference on Neural Networks (IJCNN), pp. 1–8 (2018)
DiAdamo, S., Notzel, J., Zanger, B., Beşe, M.M.: Qunetsim: a software framework for quantum network. IEEE Trans Quantum Eng. 2, 1–12 (2020)
Liu, W., Gao, P., Wang, Y., Yu, W., Zhang, M.: A unitary weights based one-iteration quantum perceptron algorithm for non-ideal training sets. IEEE Access 7, 36854–36865 (2019)
van den Berg, E.: Iterative quantum phase estimation with optimized sample complexity. In: 2020 IEEE International Conference on Quantum Computing and Engineering (QCE), pp. 1–10 (2020)
Mohammadbagherpoor, H., Oh, Y., Dreher, P., Singh, A., Yu, X., Rindos, A.J.: An improved implementation approach for quantum phase estimation on quantum computers. In: IEEE International Conference on Rebooting Computing (ICRC), pp. 1–9 (2019)
Gardner, E.: Mathematical and general. J. Phys. 21, 257 (1988)
Zhao, J.-Y.: Implementing associative memory with quantum neural networks. In: Proceedings of 2004 International Conference on Machine Learning and Cybernetics, pp. 3197–3200 (2004)
Guo, C.: Accelerating sparse DNN models without hardware-support via tile-wise sparsity. In: International Conference for High Performance Computing, Networking, Storage and Analysis, vol. 20, pp. 1–15 (2020)
Argueta, A., Chiang, D.: Accelerating sparse matrix operations in neural networks on graphics processing units. In: Accelerating Sparse Matrix Operations in Neural Networks on Graphics Processing Units, vol. 6224, no. 10, pp. 1619-1626 (2019)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 Springer Nature Switzerland AG
About this paper
Cite this paper
Ravishankar, A., Natarajan, S., Bharathi Malakreddy, A. (2022). SparseMAX: Accelerating Quantum Neural Networks on GPU Clusters Using Sparse-Matrix Kernels. In: Orailoglu, A., Jung, M., Reichenbach, M. (eds) Embedded Computer Systems: Architectures, Modeling, and Simulation. SAMOS 2021. Lecture Notes in Computer Science, vol 13227. Springer, Cham. https://doi.org/10.1007/978-3-031-04580-6_28
Download citation
DOI: https://doi.org/10.1007/978-3-031-04580-6_28
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-04579-0
Online ISBN: 978-3-031-04580-6
eBook Packages: Computer ScienceComputer Science (R0)