ABSTRACT
In recent years, node embedding algorithms, which learn low dimensional vector representations for nodes in a graph, have been one of the key research interests of the graph mining community. The existing algorithms either rely on computationally expensive eigendecomposition of the large matrices, or require tuning of the word embedding-based hyperparameters as a result of representing the graph as a node sequence similar to the sentences in a document. Moreover, the latent features produced by these algorithms are hard to interpret. In this paper, we present Tensor Decomposition-based Node Embedding (TDNE), a novel model for learning node representations for arbitrary types of graphs: undirected, directed, and/or weighted. Our model preserves the local and global structural properties of a graph by constructing a third-order tensor using the k-step transition probability matrices and decomposing the tensor through CANDECOMP/PARAFAC (CP) decomposition in order to produce an interpretable, low dimensional vector space for the nodes. Our experimental evaluation using two well-known social network datasets proves TDNE to be interpretable with respect to the understandability of the feature space, and precise with respect to the network reconstruction.
- M. Belkin and P. Niyogi. 2001. Laplacian Eigenmaps and Spectral Techniques for Embedding and Clustering. In NIPS, 2001.Google Scholar
- S. Cao, W. Lu, and Q. Xu. 2015. GraRep: Learning Graph Representations with Global Structural Information. In CIKM, 2015.Google Scholar
- P. Goyal and E. Ferrara. 2018. Graph embedding techniques, applications, and performance: A survey. Knowledge Based Systems 151 (2018), 78--94.Google ScholarCross Ref
- A. Grover and J. Leskovec. 2016. node2vec: Scalable Feature Learning for Networks. In KDD, 2016.Google ScholarDigital Library
- T. Kolda and B. Bader. 2009. Tensor decompositions and applications. SIAM review 51, 3 (2009), 455--500.Google Scholar
- T. Mikolov, K. Chen, G. Corrado, and J. Dean. 2013. Efficient Estimation of Word Representations in Vector Space. In ICLR, 2013.Google Scholar
- M. Ou, P. Cui, J. Pei, Z. Zhang, and W. Zhu. 2016. Asymmetric Transitivity Preserving Graph Embedding. In KDD, 2016.Google Scholar
- B. Perozzi, R. Al-Rfou, and S. Skiena. 2014. DeepWalk: online learning of social representations. In KDD, 2014.Google Scholar
- S. Roweis and L. Saul. 2000. Nonlinear dimensionality reduction by locally linear embedding. Science 290, 5500 (2000), 2323--2326.Google ScholarCross Ref
- N. Sidiropoulos, L. De Lathauwer, X. Fu, K. Huang, E. Papalexakis, and C. Faloutsos. 2017. Tensor decomposition for signal processing and machine learning. IEEE Trans. on Signal Processing 65, 13 (2017), 3551--3582.Google ScholarDigital Library
- W. Zachary. 1977. An information flow model for conflict and fission in small groups. Journal of anthropological research 33, 4 (1977), 452--473.Google ScholarCross Ref
- Z. Zhang, P. Cui, X. Wang, J. Pei, X. Yao, and W. Zhu. 2018. Arbitrary-Order Proximity Preserved Network Embedding. In KDD, 2018.Google Scholar
Index Terms
- Tensor Decomposition-based Node Embedding
Recommendations
Tensor Completion via Fully-Connected Tensor Network Decomposition with Regularized Factors
AbstractThe recently proposed fully-connected tensor network (FCTN) decomposition has a powerful ability to capture the low-rankness of tensors and has achieved great success in tensor completion. However, the FCTN decomposition-based method is highly ...
Node Embedding Preserving Graph Summarization
Graph summarization is a useful tool for analyzing large-scale graphs. Some works tried to preserve original node embeddings encoding rich structural information of nodes on the summary graph. However, their algorithms are designed heuristically and not ...
A tensor-based dictionary learning approach to tomographic image reconstruction
AbstractWe consider tomographic reconstruction using priors in the form of a dictionary learned from training images. The reconstruction has two stages: first we construct a tensor dictionary prior from our training data, and then we pose the ...
Comments