Elsevier

Neurocomputing

Volume 508, 7 October 2022, Pages 47-57
Neurocomputing

DynHEN: A heterogeneous network model for dynamic bipartite graph representation learning

https://doi.org/10.1016/j.neucom.2022.08.050Get rights and content

Abstract

Learning node representations in graphs is a widespread problem in node classification and link prediction. Current research has focused on static heterogeneous, homogeneous networks, and dynamic homogeneous networks. However, many existing graphs, such as citation and social networks, are heterogeneous. Therefore, it is still a great challenge to obtain heterogeneous information in a dynamic heterogeneous graph to assist node representation learning. In this paper, we propose a Dynamic HEterogeneous Network (DynHEN) for user-item bipartite networks. It is a discrete dynamic graph neural network model that can be used directly for node representation learning by utilizing dynamic heterogeneous graphs. Specifically, DynHEN takes a bipartite graph at each time step as input, gets the corresponding embedding by capturing the deep heterogeneous information of the nodes while fusing the temporal information. To illustrate the effectiveness of heterogeneous information for graph representation learning, we compare with the current SOTA methods in the two type experiments of link prediction and node classification, and achieve outstanding results.

Introduction

Representation learning of nodes in graphs has recently attracted extensive research interest and gradually evolved into a crucial problem. This technique has been widely used in various fields such as biology [1], recommendation [2], [3], [4], road condition prediction [5], and semantic analysis [6]. The graphs that people encounter in real applications change over time and have various node types, called dynamic heterogeneous graphs. For example, the continuous emergence of scientific articles and novel techniques will increase both the complexity and scale of the scientific citation network over time. Therefore, articles and scholars are always in a state of change, so the node embedding of articles and scholars needs to be updated ceaselessly. Similarly, statistics of the consumption platform show that the commodities clicked by users will continuously change with time as well as their preferences. The commodities will always emerge, and users will also keep changing. Vector representations of user preferences and product characteristics should also be updated to reflect this change accurately. These scenes widely occur in the real world. Many studies use a bipartite graph to model the above data and analyze the embedding of entities and relations, such as [7], [8]. Bipartite network is a special kind of heterogeneous graph, Fig. 1 is the bipartite network of the MovieLens dataset.1 The ubiquitous dynamic bipartite network makes the representation learning of nodes in it more and more significant. Meanwhile, the dynamics and structural heterogeneity of dynamic networks provide rich information for node embedding.

Dynamic graph neural network (DGNN) is a kind of neural network structure that encodes a dynamic network and can be considered a network where nodes and edges appear and/or disappear over time [9]. The most important part of the structure is aggregating each node’s neighbor information in the dynamic graph. DGNN is usually an elaborate combination of a recursive neural network (RNN) and a graph neural network (GNN) [10], where the GNN captures the structural information in the graph, while RNN extracts the temporal information. The DGNN can be roughly divided into the discrete version and the continuous version [9]. Discrete DGNNs usually use an independent GNN to process each snapshot, and then feed each GNN output to a time series component. At the same time, the continuous DGNNs are usually a newly created model that parameterizes the graph, and the node embeddings vary according to changes in nodes and edges. There are many discrete DGNNs, such as RgCNN [11], which uses GCN to model graph topology, and LSTM [12] (a kind of RNN) to model the time dependency, DyGGCN [13], a model organizes GGNN [14] and LSTM, and DySAT [15], which chooses graph attention networks (GAT) [16] as the GNN and substitute Self-Attention for RNN. Therefore, GNN is the core component of DGNN.

However, most of the discrete DGNNs for dynamic bipartite networks convert the bipartite networks into homogeneous graphs and use the idea of discrete DGNNs to learn node embeddings. For example, DySAT [15] makes no distinction between node types in the bipartite graph. Instead, it uses the graph attention network model to capture the graph’s structure and extends it to temporal graph snapshots of the dynamic bipartite graph. Simultaneously, EvolveGCN [17], as a discrete DGNN, also treats all nodes equally and uses RNN evolutionary GCN parameters to extract the dynamic information of graph sequence. Although these methods possess higher computing efficiency, they ignore the ample heterogeneous information of the data itself, resulting in node embedding inadequacy in the bipartite network. Plenty of studies [18], [19] have proved the importance of heterogeneous information to node embedding in graphs. Heterogeneous information can help enhance node embedding while bringing more interpretability, especially for bipartite networks. The neighbor nodes of any node in the bipartite network are other types of nodes, as shown in Fig. 1. Thus, user node embedding with aggregated neighbor information can be interpreted as user preference embedding for heterogeneous information in a bipartite graph. In contrast, item node embedding can represent a certain type of user preference. Based on this observation, we intuitively argue that the heterogeneous information in the bipartite network is meaningful for node representation in the graph.

In view of the above problems, we propose DynHEN, a node embedding technique for dynamic bipartite networks in dynamic heterogeneous graphs, which can fully capture heterogeneous information while learning node embeddings in dynamic graphs. DynHEN consists of three main parts: HGCN layer, heterogeneous GAT layer, and temporal self-attention layer. Heterogeneous graphs contain abundant structural and semantic information. Meta-path can preserve the semantic features of heterogeneous graphs to the greatest extent [20]. So we first design HGCN, which uses the meta-path to obtain the neighbor nodes of different types and depths and then uses different weights to capture information of the nodes. GAT [16] can adaptively assign weights to the neighborhood information of nodes and can capture the structure information of graphs. Based on this, we design heterogeneous GAT to capture the structure information in heterogeneous graphs. In addition, dynamic graphs usually have periodic patterns. Self-attention [21] can extract context from all past graph snapshots to adaptively assign interpretable weights to previous time steps, so we use it to capture the dynamic changes of node features.

Then, the process of DynHEN is as follows. First, DynHEN selects an appropriate time step to represent the dynamic bipartite network as multiple bipartite network snapshots with this time step interval. Then, each snapshot uses the heterogeneous graph convolutional neural network (HGCN) to learn different types of node embeddings. Next, heterogeneous GAT is used to extract heterogeneous information further. Finally, we use the temporal self-attention to learn the representations of different nodes under each time step. In this way, we can capture the heterogeneous and dynamic information of the dynamic bipartite network. We perform extensive experiments on two tasks: link prediction and node classification, and verify the effectiveness of our proposed method. Our contributions can be summarized as follows:

  • We propose DynHEN, a heterogeneous graph neural network model for dynamic bipartite networks belonging to discrete DGNNs, to utilize the heterogeneous information in bipartite graphs effectively. The model can comprehensively capture heterogeneous information in dynamic bipartite networks.

  • We organically combine temporal information to learn dynamic node heterogeneous embedding when extracting heterogeneous information.

  • We conduct link prediction experiments on three real datasets with different sparsity, and do node classification experiments on two publicly available datasets, which proves the effectiveness of our model.

The rest of this paper is structured as follows. In Section 2, we introduce some work related to our proposed model. In Section 3, we introduce our model in detail. In Section 4, we describe the design of experiments and the analysis of results. Section 5 is the summary of this paper.

Section snippets

Related work

We describe the related work of dynamic graph representation learning including static graph embedding and dynamic graph embedding.

Proposed method

In this section, we describe our proposed method (DynHEN). We first give some necessary definitions and then provide an overview of the model architecture with a detailed description afterward.

Experiment

We refer to the experimental design of [15], take the link prediction experiment as the main experiment to analyze the performance of DynHEN fully, and take the node classification experiment as the auxiliary experiment to further prove the advantages of DynHEN in bipartite graph representation learning. The code is written in PyTorch and runs on NVIDIA RTX 3090.

Conclusion

This paper proposes a dynamic heterogeneous graph neural network model named DynHEN to learn node embeddings in dynamic bipartite graphs. This method uses heterogeneous graph convolutional neural network and the heterogeneous graph attention network to capture heterogeneous information in bipartite graphs to enhance node representation learning. Finally, the temporal sequence information is fused with the multi-head self-attention method to learn node embedding in a dynamic bipartite graph. Our

CRediT authorship contribution statement

Zhezhe Xing: Conceptualization, Methodology, Data curation, Writing – original draft. Rui Song: Writing – review & editing. Yun Teng: Visualization. Hao Xu: Supervision.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Acknowledgments

This research is supported by the National Natural Science Foundation of China (62077027), the Ministry of Science and Technology of the People’s Republic of China(2018YFC2002500), the Jilin Province Development and Reform Commission, China (2019C053-1), the Education Department of Jilin Province, China (JJKH20200993K), the Department of Science and Technology of Jilin Province, China (20200801002GH), and the European Union’s Horizon 2020 FET Proactive project ”WeNet-The Internet of us” (No.

Zhezhe Xing is currently studying for a Master’s degree in the Software College, Jilin University. Her main research interests include graph neural networks and recommendation system.

References (42)

  • A. Narayan et al.

    Learning graph dynamics using deep neural networks - sciencedirect

    IFAC-PapersOnLine

    (2018)
  • W.L. Hamilton, Z. Ying, J. Leskovec, Inductive representation learning on large graphs, in: Advances in Neural...
  • C. Shi et al.

    Heterogeneous information network embedding for recommendation

    IEEE Trans. Knowl. Data Eng.

    (2019)
  • M. Zhang, S. Wu, X. Yu, L. Wang, Dynamic graph neural networks for sequential recommendation, CoRR abs/2104.07368....
  • C. Li, Z. Liu, M. Wu, Y. Xu, H. Zhao, P. Huang, G. Kang, Q. Chen, W. Li, D.L. Lee, Multi-interest network with dynamic...
  • B. Yu, H. Yin, Z. Zhu, Spatio-temporal graph convolutional neural network: A deep learning framework for traffic...
  • D. Marcheggiani, I. Titov, Encoding sentences with graph convolutional networks for semantic role labeling, in:...
  • Y. Cen, X. Zou, J. Zhang, H. Yang, J. Zhou, J. Tang, Representation learning for attributed multiplex heterogeneous...
  • B. Kaya, Hotel recommendation system by bipartite networks and link prediction, J. Inf. Sci. 46 (1)....
  • J. Skarding, B. Gabrys, K. Musial, Foundations and modelling of dynamic networks using dynamic graph neural networks: A...
  • F. Scarselli et al.

    The graph neural network model

    IEEE Trans. Neural Networks

    (2009)
  • S. Hochreiter et al.

    Long short-term memory

    Neural Comput.

    (1997)
  • A. Taheri, K. Gimpel, T.Y. Berger-Wolf, Learning to represent the evolution of dynamic graphs with recurrent models,...
  • Y. Li, D. Tarlow, M. Brockschmidt, R.S. Zemel, Gated graph sequence neural networks, in: 4th International Conference...
  • A. Sankar, Y. Wu, L. Gou, W. Zhang, H. Yang, Dysat: Deep neural representation learning on dynamic graphs via...
  • P. Velickovic, G. Cucurull, A. Casanova, A. Romero, P. Liò, Y. Bengio, Graph attention networks, in: 6th International...
  • A. Pareja, G. Domeniconi, J. Chen, T. Ma, T. Suzumura, H. Kanezashi, T. Kaler, T.B. Schardl, C.E. Leiserson, Evolvegcn:...
  • X. Sha, Z. Sun, J. Zhang, Attentive knowledge graph embedding for personalized recommendation, CoRR abs/1910.08288....
  • J. Zhao, Z. Zhou, Z. Guan, W. Zhao, W. Ning, G. Qiu, X. He, Intentgc: A scalable graph convolution framework fusing...
  • Y. Dong et al.

    metapath2vec, Scalable representation learning for heterogeneous networks

  • A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A.N. Gomez, L. Kaiser, I. Polosukhin, Attention is all you...
  • Cited by (3)

    • The Evaluation of DyHATR Performance for Dynamic Heterogeneous Graphs

      2023, Journal of ICT Research and Applications

    Zhezhe Xing is currently studying for a Master’s degree in the Software College, Jilin University. Her main research interests include graph neural networks and recommendation system.

    Rui Song is currently studying for a Ph.D. degree in the School of Artificial Intelligence, Jilin University. He received the B.S. degree from the College of Software of Jilin University. His main research interests include natural language processing and graph neural networks.

    Yun Teng received his Bachelor’s degree from Jilin University, Changchun, China, in 2015. He is pursuing his Master’s degree candidate in the School of Computer Science and Technology, Jilin University, Changchun, China. His research interests include distributed storage and disk arrays.

    Hao Xu received the B.S. and M.S. degrees in computer science from Jilin University, Changchun, China, in 2005 and 2008, respectively, and the Ph.D. degree from University of Trento, Trento, Italy, in 2011. He is a Professor with Jilin University. His research interests include human-centered artificial intelligence, knowledge representation and graph neural networks.

    View full text