Abstract
Link prediction aims to capture the evolution of network structure, especially in real social networks, which is conducive to friend recommendations, human contact trajectory simulation, and more. However, the challenge of the stochastic social behaviors and the unstable space-time distribution in such networks often leads to unexplainable and inaccurate link predictions. Therefore, taking inspiration from the success of imitation learning in simulating human driver behavior, we propose a dynamic network link prediction method based on inverse reinforcement learning (DN-IRL) to unravel the motivations behind social behaviors in social networks. Specifically, the historical social behaviors (link sequences) and a next behavior (a single link) are regarded as the current environmental state and the action taken by the agent, respectively. Subsequently, the reward function, which is designed to maximize the cumulative expected reward from expert behaviors in the raw data, is optimized and utilized to learn the agent’s social policy. Furthermore, our approach incorporates the neighborhood structure based node embedding and the self-attention modules, enabling sensitivity to network structure and traceability to predicted links. Experimental results on real-world dynamic social networks demonstrate that DN-IRL achieves more accurate and explainable of prediction compared to the baselines.










Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data availibility statement
The data in this work comes from an open network repository [35], and can be get in this web link (https://networkrepository.com/dynamic.php).
References
Du W, Li G, He X (2022) Network structure optimization for social networks by minimizing the average path length. Computing 104(6):1461–1480. https://doi.org/10.1007/s00607-022-01061-w
Kumari A, Behera RK, Sahoo B et al (2022) Prediction of link evolution using community detection in social network. Computing 104(5):1077–1098. https://doi.org/10.1007/s00607-021-01035-4
Flores-Martin D, Berrocal J, García-Alonso J et al (2023) Towards dynamic and heterogeneous social IoT environments. Computing 105(6):1141–1164. https://doi.org/10.1007/s00607-022-01113-1
Chen J, Xu X, Wu Y, et al (2018) GC-LSTM: graph convolution embedded LSTM for dynamic link prediction. https://doi.org/10.48550/arXiv.1812.04206, arXiv:1812.04206
Chen J, Zhang J, Xu X et al (2021) E-LSTM-D: a deep learning framework for dynamic network link prediction. IEEE Trans Syst Man Cybern Syst 51(6):3699–3712
La Gatta V, Moscato V, Postiglione M et al (2021) An epidemiological neural network exploiting dynamic graph structured data applied to the covid-19 outbreak. IEEE Trans Big Data 7(1):45–55
Yang M, Liu J, Chen L et al (2020) An advanced deep generative framework for temporal link prediction in dynamic networks. IEEE Trans Cybern 50(12):4946–4957
Feng F, He X, Tang J et al (2021) Graph adversarial training: dynamically regularizing based on graph structure. IEEE Trans Knowl Data Eng 33(6):2493–2504
Sandryhaila A, Moura JMF (2014) Discrete signal processing on graphs: frequency analysis. IEEE Trans Signal Process 62(12):3042–3054
Yang L, Jiang X, Ji Y et al (2022) Gated graph convolutional network based on spatio-temporal semi-variogram for link prediction in dynamic complex network. Neurocomputing 505:289–303
Wu Z, Pan S, Chen F et al (2021) A comprehensive survey on graph neural networks. IEEE Trans Neural Netw Learn Syst 32(1):4–24
Li Q, Shen B, Wang Z et al (2019) Synchronization control for a class of discrete time-delay complex dynamical networks: a dynamic event-triggered approach. IEEE Trans Cybern 49(5):1979–1986
Li K, Rath M, Burdick JW (2018) Inverse reinforcement learning via function approximation for clinical motion analysis. In: 2018 IEEE international conference on robotics and automation (ICRA), pp 610–617
Naumann M, Sun L, Zhan W, et al (2020) Analyzing the suitability of cost functions for explaining and imitating human driving behavior based on inverse reinforcement learning. In: 2020 IEEE international conference on robotics and automation (ICRA), pp 5481–5487
Wu Z, Sun L, Zhan W et al (2020) Efficient sampling-based maximum entropy inverse reinforcement learning with application to autonomous driving. IEEE Robot Autom Lett 5(4):5355–5362
Hamilton WL, Ying R, Leskovec J (2017) Inductive representation learning on large graphs. In: Proceedings of the 31st international conference on neural information processing systems. Curran Associates Inc., Red Hook, NY, USA, NIPS’17, pp 1025–1035
Zhao Z, Lin S (2023) A cross-linguistic entity alignment method based on graph convolutional neural network and graph attention network. Computing. https://doi.org/10.1007/s00607-023-01178-6
Wang J, Liang J, Yao K et al (2022) Graph convolutional autoencoders with co-learning of graph structure and node attributes. Pattern Recogn 121:108215. https://doi.org/10.1016/j.patcog.2021.108215
Pareja A, Domeniconi G, Chen J, et al (2020) EvolveGCN: evolving graph convolutional networks for dynamic graphs. In: Proceedings of the 34th AAAI conference on artificial intelligence. AAAI Press, Palo Alto, CA, pp 5679–5681
Du L, Wang Y, Song G, et al (2018a) Dynamic network embedding: an extended approach for skip-gram based network embedding. In: Proceedings of the 27th AAAI conference on artificial intelligence. AAAI Press, Palo Alto, CA, IJCAI’18, pp 2086–2092
Du L, Wang Y, Song G, et al (2018b) Dynamic network embedding: an extended approach for skip-gram based network embedding. In: Proceedings of the 27th international joint conference on artificial intelligence. AAAI Press, IJCAI’18, pp 2086–2092
Hou C, Zhang H, He S et al (2022) GloDyNE: global topology preserving dynamic network embedding. IEEE Trans Knowl Data Eng 34(10):4826–4837
Xu D, Ruan C, Korpeoglu E, et al (2020) Inductive representation learning on temporal graphs. In: Proceedings of international conference on learning representations. OpenReview.net, Ithaca, NY, https://openreview.net/forum?id=rJeW1yHYwH
Jiang X, Yu Z, Hai C et al (2023) DNformer: temporal link prediction with transfer learning in dynamic networks. ACM Trans Knowl Discove Data. https://doi.org/10.1145/3551892
Xu K, Hu W, Leskovec J, et al (2019) How powerful are graph neural networks? In: International conference on learning representations https://doi.org/10.48550/arXiv.1810.00826
Kalman RE (1964) When is a linear control system optimal? J Basic Eng 86(1):51–60. https://doi.org/10.1115/1.3653115
Ng AY, Russell SJ (2000) Algorithms for inverse reinforcement learning. In: Proceedings of the seventeenth international conference on machine learning. Morgan Kaufmann Publishers Inc., San Francisco, CA, USA, ICML ’00, pp 663–670
You C, Lu J, Filev D et al (2019) Advanced planning for autonomous vehicles using reinforcement learning and deep inverse reinforcement learning. Robot Auton Syst 114:1–18. https://doi.org/10.1016/j.robot.2019.01.003
Ziebart BD, Maas AL, Dey AK, et al (2008a) Navigate like a cabbie: probabilistic reasoning from observed context-aware behavior. In: Proceedings of the 10th international conference on ubiquitous computing. Association for Computing Machinery, New York, UbiComp ’08, pp 322–331
Ziebart BD, Maas AL, Bagnell JA, et al (2008b) Maximum entropy inverse reinforcement learning. In: Proceedings of the 23rd AAAI conference on artificial intelligence. Association for the Advancement of Artificial Intelligence, New York, pp 1433–1438
Morgenstern O, Von Neumann J (1953) Theory of games and economic behavior. Princeton University Press, Princeton
So W, Robbiano M, de Abreu NMM et al (2010) Applications of a theorem by Ky Fan in the theory of graph energy. Linear Algebra Appl 432(9):2163–2169. https://doi.org/10.1016/j.laa.2009.01.006
Liu CH, Dai Z, Zhao Y et al (2021) Distributed and energy-efficient mobile crowdsensing with charging stations by deep reinforcement learning. IEEE Trans Mob Comput 20(1):130–146. https://doi.org/10.1109/TMC.2019.2938509
Vaswani A, Shazeer N, Parmar N, et al (2017) Attention is all you need. In: Proceedings of the 31st international conference on neural information processing systems. Curran Associates Inc., Red Hook, NIPS’17, pp 6000–6010
Rossi RA, Ahmed NK (2015) The network data repository with interactive graph analytics and visualization. In: Bonet B, Koenig S (eds) Proceedings of the 29th AAAI conference on artificial intelligence, January 25-30, 2015, Austin, Texas. AAAI Press, pp 4292–4293
Adler J, Lunz S (2018) Banach wasserstein gan. In: Proceedings of the 32nd international conference on neural information processing systems. Curran Associates Inc., Red Hook, NIPS’18, pp 6755–6764
Junuthula RR, Xu KS, Devabhaktuni VK (2016) Evaluating link prediction accuracy in dynamic networks with added and removed edges. In: 2016 IEEE international conferences on big data and cloud computing (BDCloud), Social Computing and Networking (SocialCom), Sustainable Computing and Communications (SustainCom) (BDCloud-SocialCom-SustainCom), pp 377–384
Pu C, Li J, Wang J et al (2022) The node-similarity distribution of complex networks and its applications in link prediction. IEEE Trans Knowl Data Eng 34(8):4011–4023
Perozzi B, Al-Rfou R, Skiena S (2014) DeepWalk: online learning of social representations. In: Proceedings of the 20th ACM SIGKDD international conference on knowledge discovery and data mining. Association for Computing Machinery, New York, pp 701–710
Acknowledgements
This work was supported by two National Natural Science Foundation of China under Grant 61772102 and 62176036, and by the Science Foundation Ireland under grant SFI/12/RC/2289_P2.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
All authors certify that they have no affiliations with or involvement in any organization or entity with any financial interest or non-financial interest in the subject matter or materials discussed in this manuscript.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Proof of Theorem
Proof of Theorem
The neighborhood structure vector maps the neighborhood information of each node included in the adjacency matrix to the vector space with fixed dimensions. In this section, we prove the injectivity of the mapping.
Theorem 1
(Injective mapping of node neighborhood structure) Given any one graph \(G=(V,E)\) and its adjacency matrix A, the neighborhood topological vector \(X_i=[x_{i1},x_{i2},\ldots ,x_{il}]\) of node \(v_i\) are obtained by summing the matrices \(\{A,A^2,\ldots ,A^l\}\) by row. If and only if the neighborhood structures of node \(v_i\) are isomorphic as the node \(v_j\), their neighborhood topological vectors are the same \(X_i=X_j\).
Proof
Firstly, the element \(x_{il}\) represents the number of different walks with length l from the node i. If \(l=2\), \(x_{i2}\) is equal to
where \(a_{ik}^{_{(1)}}\) and \(a_{im}^{_{(2)}}\) are the values of A[i, j] and \(A^2[i,j]\), respectively. If only if \(a_{ik}^{_{(1)}}=1\) and \(a_{km}^{_{(1)}}=1\), \(a_{im}^{_{(2)}}=1\), namely, there exists a walk from the node \(v_i\) to node \(v_m\), \((i\rightarrow k\rightarrow m)\) in the graph G. Therefore, \(x_{i2}\) denotes the total number of walks with length 2 from node i to arbitrary nodes. Clearly, the value of \(X_i\) is solely determined by the neighborhood structure of node \(v_i\). Nodes with symmetrical structures are inherently guaranteed to possess identical neighborhood feature vectors.
Secondly, we emphasize the rationale behind the distinct neighborhood feature vectors corresponding to different node neighborhood structures. Two typical case of the distinct node neighborhood structures are shown in Fig. 11.
Obviously in the Case 1, the discrepancy in the number of k-th hop neighboring nodes between the nodes 1 and a results in distinct values for \(x_{1k}\) and \(x_{ak}\). Therefore, we need to focus on the Case 2 where the number of k-th hop neighbors is the same for two nodes, but the neighborhood structures are different.
To formalize the neighborhood structure vector, we assume \(d_0\), \(d_1\) and \(d_2\) represent the root node \(v_i\), its neighbor nodes and its 2-hop nodes, respectively. Then, the value of \(x_{i4}\) in the neighborhood topological vector \(X_i\) is
where \((d_0\rightarrow d_1\rightarrow d_2\rightarrow d_1\rightarrow d_2)\) means all the walks from the root node \(v_i\) to the 2-hop nodes when the length of walks is equal to 4. We can observe that the number of walks \((d_0\rightarrow d_1\rightarrow d_2\rightarrow d_1\rightarrow d_2)\) is related to the allocation of 2-hop neighbor nodes. The reason of \(x_{14}\ne x_{a4}\) is
Furthermore, we suppose \(|d_1|=n,|d_2|=m\) represent the number of neighbor nodes and 2-hop nodes of the node \(v_i\), respectively. And the allocation of 2-hop nodes is
where \(k_s\) represents the number of 2-hop nodes connected with the s-th neighbor node. Then, Eq. (17) is changed into
where \(\sum _{s=1}^{n}k_s^2\) is the number of walks \((d_0\rightarrow d_1\rightarrow d_2\rightarrow d_1\rightarrow d_2)\), in which both \(d_1^{_{(s)}}\rightarrow d_2\) and \(d_2\rightarrow d_1^{_{(s)}}\rightarrow d_2\) have \(k_s\) walks, \(d_1^{_{(s)}}\in d_1\). For example, \(x_{a4}=(2^2+0^2+2^2)+(3\times 4+3^2+4)=33\), \(x_{14}=(2^2+1^2+1^2)+(3\times 4+3^2+4)=31\). From Eq. (20), when the nodes \(v_i\) and \(v_j\) have the same number of 2-hop neighbors but different allocation methods, the necessary condition of \(x_{i4}=x_{j4}\) is,
However, Eq. (21) does have a positive integer solution, for example, if \(m=6\), \(\{k_{i1},k_{i2},k_{i3},k_{i4}\}=\{2,2,2,0\}\) and \(\{k_{j1},k_{j2},k_{j3},k_{j4}\}=\{3,1,1,1\}\). Thus we must consider setting a larger l, such as the value of \(x_{i6}\) is
where O(n, m) represents the polynomial containing parameters n and m. The reason for the occurrence of the third power of \(\{k_s\}\) in Eq. (22) is due to the special walk shown in Fig. 12. Simply put, there are \(k_s\) possibilities each time a random walk moves from the s-th neighboring node to its 2-hop nodes.
Then, when \(\{k_{i1},k_{i2},\ldots ,k_{in}\}\ne \{k_{j1},k_{j2},\ldots ,k_{jn}\}\), \(l=2N\), \(|d_1|=n\), \(|k_i|=|k_j|=m\) and \(X_i=X_j\), the following equations must be met:
where \(k_{is},k_{js}\subset \mathbf {N^*}\), and Eq. (23) can be expressed as an integer programming problem. More importantly, n and m are are two specific positive integers, both of which are smaller than the total number of nodes in the entire network (typically in the range of hundreds to thousands). Therefore, according to the combination theory, the number of solutions in the feasible solution space is finite, and it is
Furthermore, as the value of l increases, it indicates a higher number of constraints in the integer problem of Eq. (23), resulting in a lower probability of finding feasible solutions. In other words, there is a injective mapping from the neighborhood structure to the neighborhood topological vector \(\{x_{i1},x_{i2},\ldots ,x_{in}\}\), when l is set as a sufficiently large integer. \(\square\)
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Jiang, X., Liu, H., Yang, L. et al. Unraveling human social behavior motivations via inverse reinforcement learning-based link prediction. Computing 106, 1963–1986 (2024). https://doi.org/10.1007/s00607-024-01279-w
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00607-024-01279-w