Learning Low-Rank Structured Sparsity in Recurrent Neural Networks | IEEE Conference Publication | IEEE Xplore

Learning Low-Rank Structured Sparsity in Recurrent Neural Networks


Abstract:

Acceleration and wide deployability in deeper recurrent neural network is hindered by high demand for computation and memory storage on devices with memory and latency co...Show More

Abstract:

Acceleration and wide deployability in deeper recurrent neural network is hindered by high demand for computation and memory storage on devices with memory and latency constraints. In this work, we propose a novel regularization method to learn hardware-friendly sparse structures for deep recurrent neural networks. Considering the consistency of dimension in continuous time units in recurrent neural networks, low-rank structured sparse approximations of the weight matrices are learned through the regularization without dimension distortion. Our method is architecture agnostic and can learn compact models with higher degree of sparsity than the state-of-the-art structured sparsity learning method. The structured sparsity rather than random sparsity also facilitates the hardware implementation. Experiments on language modeling of Penn TreeBank dataset show that our approach can reduce the parameters of stacked recurrent neural network model by over 90% with less than 1% perplexity loss. It is also successfully evaluated on larger highway neural network model with word2vec dataset like enwik8 and text8 using only 20M weights.
Date of Conference: 12-14 October 2020
Date Added to IEEE Xplore: 28 September 2020
Print ISBN:978-1-7281-3320-1
Print ISSN: 2158-1525
Conference Location: Seville, Spain

Contact IEEE to Subscribe

References

References is not available for this document.