ABSTRACT
Large scale high dimensional multi-aspect data are naturally represented as tensors or multi-way arrays for scientific and engineering computations. But in practical applications such as signal processing, data mining, computer vision, and graph analysis, this tensor data is very sparse. The degree of sparsity is magnified by the increase of number of dimensions for large tensors. Therefore, storing and applying tensor operations on this highly sparse multidimensional data is an important research challenge for data scientists. In this paper, we describe a new sparse tensor storage format that provide storage benefits and are independent to the number of dimension of the tensor. Efficient hash functions are desined to store sparse tensor data. The hash functions convert a higher order tensor to a matrix by tensor unfolding. And using the unfolded matrix, we develop algorithms to store nonzero data based on sparse fibers. We call our scheme Unfolded Compressed Row/Column Fiber (UCRF/UCCF). Our approach experimental result shows superior performance with standard dataset comparing to other important approaches.
- [n. d.]. Formidable Repository of Open Sparse Tensors and Tools. http://frostt.io/. Accessed: 2021-03-13.Google Scholar
- Brett W Bader and Tamara G Kolda. 2008. Efficient MATLAB computations with sparse and factored tensors. SIAM Journal on Scientific Computing 30, 1 (2008), 205--231.Google ScholarDigital Library
- Zhaojun Bai, James Demmel, Jack Dongarra, Axel Ruhe, and Henk van der Vorst. 2000. Templates for the solution of algebraic eigenvalue problems: a practical guide. SIAM.Google Scholar
- Richard Barrett, Michael Berry, Tony F Chan, James Demmel, June Donato, Jack Dongarra, Victor Eijkhout, Roldan Pozo, Charles Romine, and Henk Van der Vorst. 1994. Templates for the solution of linear systems: building blocks for iterative methods. SIAM.Google Scholar
- James Bennett, Stan Lanning, et al. 2007. The netflix prize. In Proceedings of KDD cup and workshop, Vol. 2007. New York, NY, USA., 35.Google Scholar
- Aydin Buluç, Jeremy T Fineman, Matteo Frigo, John R Gilbert, and Charles E Leiserson. 2009. Parallel sparse matrix-vector and matrix-transpose-vector multiplication using compressed sparse blocks. In Proceedings of the twenty-first annual symposium on Parallelism in algorithms and architectures. 233--244.Google ScholarDigital Library
- Aydin Buluc and John R Gilbert. 2008. On the representation and multiplication of hypersparse matrices. In 2008 IEEE International Symposium on Parallel and Distributed Processing. IEEE, 1--11.Google ScholarCross Ref
- Andrew Carlson, Justin Betteridge, Bryan Kisiel, Burr Settles, Estevam R Hruschka, and Tom M Mitchell. 2010. Toward an architecture for never-ending language learning. In Twenty-Fourth AAAI conference on artificial intelligence.Google ScholarCross Ref
- Joon Hee Choi and S Vishwanathan. 2014. DFacTo: Distributed factorization of tensors. Advances in Neural Information Processing Systems 27 (2014), 1296--1304.Google Scholar
- Ryan Eberhardt and Mark Hoemmen. 2016. Optimization of block sparse matrix-vector multiplication on shared-memory parallel architectures. In 2016 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW). IEEE, 663--672.Google ScholarCross Ref
- Fredrik Kjolstad, Shoaib Kamil, Stephen Chou, David Lugato, and Saman Amarasinghe. 2017. The tensor algebra compiler. Proceedings of the ACM on Programming Languages 1, OOPSLA (2017), 1--29.Google ScholarDigital Library
- Tamara Gibson Kolda. 2006. Multilinear operators for higher-order decompositions. Technical Report. Citeseer.Google Scholar
- Daniel Langr and Pavel Tvrdik. 2015. Evaluation criteria for sparse matrix storage formats. IEEE Transactions on parallel and distributed systems 27, 2 (2015), 428--440.Google ScholarDigital Library
- James Li, Jacob Bien, and Martin T Wells. 2018. rTensor: An R package for multidimensional array (tensor) unfolding, multiplication, and decomposition. Journal of Statistical Software 87, 1 (2018), 1--31.Google ScholarCross Ref
- Julian McAuley and Jure Leskovec. 2013. Hidden factors and hidden topics: understanding rating dimensions with review text. In Proceedings of the 7th ACM conference on Recommender systems. 165--172.Google ScholarDigital Library
- Niranjay Ravindran, Nicholas D Sidiropoulos, Shaden Smith, and George Karypis. 2014. Memory-efficient parallel computation of tensor and matrix products for big tensor decomposition. In 2014 48th Asilomar Conference on Signals, Systems and Computers. IEEE, 581--585.Google ScholarCross Ref
- Shaden Smith and George Karypis. 2015. Tensor-matrix products with a compressed sparse tensor. In Proceedings of the 5th Workshop on Irregular Applications: Architectures and Algorithms. 1--7.Google ScholarDigital Library
- Shaden Smith, Jongsoo Park, and George Karypis. 2017. Sparse tensor factorization on many-core processors with high-bandwidth memory. In 2017 IEEE International Parallel and Distributed Processing Symposium (IPDPS). IEEE, 1058--1067.Google ScholarCross Ref
- Shaden Smith, Niranjay Ravindran, Nicholas D Sidiropoulos, and George Karypis. 2015. SPLATT: Efficient and parallel sparse tensor-matrix multiplication. In 2015 IEEE International Parallel and Distributed Processing Symposium. IEEE, 61--70.Google ScholarDigital Library
- George Strawn. 2014. Don Knuth: Mastermind of Algorithms [review of" The art of programming"]. IT Professional 16, 5 (2014), 70--72.Google ScholarCross Ref
- Bimal Viswanath, Alan Mislove, Meeyoung Cha, and Krishna P Gummadi. 2009. On the evolution of user interaction in facebook. In Proceedings of the 2nd ACM workshop on Online social networks. 37--42.Google ScholarDigital Library
- James B White and Ponnuswamy Sadayappan. 1997. On improving the performance of sparse matrix-vector multiplication. In Proceedings Fourth International Conference on High-Performance Computing. IEEE, 66--71.Google ScholarCross Ref
Index Terms
- Sparse tensor storage by tensor unfolding
Recommendations
Efficient and effective sparse tensor reordering
ICS '19: Proceedings of the ACM International Conference on SupercomputingThis paper formalizes the problem of reordering a sparse tensor to improve the spatial and temporal locality of operations with it, and proposes two reordering algorithms for this problem, which we call BFS-MCS and Lexi-Order. The BFS-MCS method is a ...
Sparse and Truncated Nuclear Norm Based Tensor Completion
One of the main difficulties in tensor completion is the calculation of the tensor rank. Recently a tensor nuclear norm, which is equal to the weighted sum of matrix nuclear norms of all unfoldings of the tensor, was proposed to address this issue. ...
Robust low tubal rank tensor completion via factor tensor norm minimization
Highlights- We give the definitions of tensor double norm and tensor Frobenius/nuclear hybrid norm, and regard them as low-rank regularization penalty of tensor ...
AbstractRecent research has demonstrated that low tubal rank recovery based on tensor has received extensive attention. In this correspondence, we define tensor double nuclear norm and tensor Frobenius/nuclear hybrid norm to induce a surrogate ...
Comments