Skip to main content

Advertisement

Log in

Unraveling human social behavior motivations via inverse reinforcement learning-based link prediction

  • Regular Paper
  • Published:
Computing Aims and scope Submit manuscript

Abstract

Link prediction aims to capture the evolution of network structure, especially in real social networks, which is conducive to friend recommendations, human contact trajectory simulation, and more. However, the challenge of the stochastic social behaviors and the unstable space-time distribution in such networks often leads to unexplainable and inaccurate link predictions. Therefore, taking inspiration from the success of imitation learning in simulating human driver behavior, we propose a dynamic network link prediction method based on inverse reinforcement learning (DN-IRL) to unravel the motivations behind social behaviors in social networks. Specifically, the historical social behaviors (link sequences) and a next behavior (a single link) are regarded as the current environmental state and the action taken by the agent, respectively. Subsequently, the reward function, which is designed to maximize the cumulative expected reward from expert behaviors in the raw data, is optimized and utilized to learn the agent’s social policy. Furthermore, our approach incorporates the neighborhood structure based node embedding and the self-attention modules, enabling sensitivity to network structure and traceability to predicted links. Experimental results on real-world dynamic social networks demonstrate that DN-IRL achieves more accurate and explainable of prediction compared to the baselines.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

Data availibility statement

The data in this work comes from an open network repository [35], and can be get in this web link (https://networkrepository.com/dynamic.php).

References

  1. Du W, Li G, He X (2022) Network structure optimization for social networks by minimizing the average path length. Computing 104(6):1461–1480. https://doi.org/10.1007/s00607-022-01061-w

    Article  MathSciNet  Google Scholar 

  2. Kumari A, Behera RK, Sahoo B et al (2022) Prediction of link evolution using community detection in social network. Computing 104(5):1077–1098. https://doi.org/10.1007/s00607-021-01035-4

    Article  Google Scholar 

  3. Flores-Martin D, Berrocal J, García-Alonso J et al (2023) Towards dynamic and heterogeneous social IoT environments. Computing 105(6):1141–1164. https://doi.org/10.1007/s00607-022-01113-1

    Article  Google Scholar 

  4. Chen J, Xu X, Wu Y, et al (2018) GC-LSTM: graph convolution embedded LSTM for dynamic link prediction. https://doi.org/10.48550/arXiv.1812.04206, arXiv:1812.04206

  5. Chen J, Zhang J, Xu X et al (2021) E-LSTM-D: a deep learning framework for dynamic network link prediction. IEEE Trans Syst Man Cybern Syst 51(6):3699–3712

    Article  Google Scholar 

  6. La Gatta V, Moscato V, Postiglione M et al (2021) An epidemiological neural network exploiting dynamic graph structured data applied to the covid-19 outbreak. IEEE Trans Big Data 7(1):45–55

    Article  Google Scholar 

  7. Yang M, Liu J, Chen L et al (2020) An advanced deep generative framework for temporal link prediction in dynamic networks. IEEE Trans Cybern 50(12):4946–4957

    Article  Google Scholar 

  8. Feng F, He X, Tang J et al (2021) Graph adversarial training: dynamically regularizing based on graph structure. IEEE Trans Knowl Data Eng 33(6):2493–2504

    Article  Google Scholar 

  9. Sandryhaila A, Moura JMF (2014) Discrete signal processing on graphs: frequency analysis. IEEE Trans Signal Process 62(12):3042–3054

    Article  MathSciNet  Google Scholar 

  10. Yang L, Jiang X, Ji Y et al (2022) Gated graph convolutional network based on spatio-temporal semi-variogram for link prediction in dynamic complex network. Neurocomputing 505:289–303

    Article  Google Scholar 

  11. Wu Z, Pan S, Chen F et al (2021) A comprehensive survey on graph neural networks. IEEE Trans Neural Netw Learn Syst 32(1):4–24

    Article  MathSciNet  Google Scholar 

  12. Li Q, Shen B, Wang Z et al (2019) Synchronization control for a class of discrete time-delay complex dynamical networks: a dynamic event-triggered approach. IEEE Trans Cybern 49(5):1979–1986

    Article  Google Scholar 

  13. Li K, Rath M, Burdick JW (2018) Inverse reinforcement learning via function approximation for clinical motion analysis. In: 2018 IEEE international conference on robotics and automation (ICRA), pp 610–617

  14. Naumann M, Sun L, Zhan W, et al (2020) Analyzing the suitability of cost functions for explaining and imitating human driving behavior based on inverse reinforcement learning. In: 2020 IEEE international conference on robotics and automation (ICRA), pp 5481–5487

  15. Wu Z, Sun L, Zhan W et al (2020) Efficient sampling-based maximum entropy inverse reinforcement learning with application to autonomous driving. IEEE Robot Autom Lett 5(4):5355–5362

    Article  Google Scholar 

  16. Hamilton WL, Ying R, Leskovec J (2017) Inductive representation learning on large graphs. In: Proceedings of the 31st international conference on neural information processing systems. Curran Associates Inc., Red Hook, NY, USA, NIPS’17, pp 1025–1035

  17. Zhao Z, Lin S (2023) A cross-linguistic entity alignment method based on graph convolutional neural network and graph attention network. Computing. https://doi.org/10.1007/s00607-023-01178-6

    Article  Google Scholar 

  18. Wang J, Liang J, Yao K et al (2022) Graph convolutional autoencoders with co-learning of graph structure and node attributes. Pattern Recogn 121:108215. https://doi.org/10.1016/j.patcog.2021.108215

    Article  Google Scholar 

  19. Pareja A, Domeniconi G, Chen J, et al (2020) EvolveGCN: evolving graph convolutional networks for dynamic graphs. In: Proceedings of the 34th AAAI conference on artificial intelligence. AAAI Press, Palo Alto, CA, pp 5679–5681

  20. Du L, Wang Y, Song G, et al (2018a) Dynamic network embedding: an extended approach for skip-gram based network embedding. In: Proceedings of the 27th AAAI conference on artificial intelligence. AAAI Press, Palo Alto, CA, IJCAI’18, pp 2086–2092

  21. Du L, Wang Y, Song G, et al (2018b) Dynamic network embedding: an extended approach for skip-gram based network embedding. In: Proceedings of the 27th international joint conference on artificial intelligence. AAAI Press, IJCAI’18, pp 2086–2092

  22. Hou C, Zhang H, He S et al (2022) GloDyNE: global topology preserving dynamic network embedding. IEEE Trans Knowl Data Eng 34(10):4826–4837

    Article  Google Scholar 

  23. Xu D, Ruan C, Korpeoglu E, et al (2020) Inductive representation learning on temporal graphs. In: Proceedings of international conference on learning representations. OpenReview.net, Ithaca, NY, https://openreview.net/forum?id=rJeW1yHYwH

  24. Jiang X, Yu Z, Hai C et al (2023) DNformer: temporal link prediction with transfer learning in dynamic networks. ACM Trans Knowl Discove Data. https://doi.org/10.1145/3551892

    Article  Google Scholar 

  25. Xu K, Hu W, Leskovec J, et al (2019) How powerful are graph neural networks? In: International conference on learning representations https://doi.org/10.48550/arXiv.1810.00826

  26. Kalman RE (1964) When is a linear control system optimal? J Basic Eng 86(1):51–60. https://doi.org/10.1115/1.3653115

    Article  Google Scholar 

  27. Ng AY, Russell SJ (2000) Algorithms for inverse reinforcement learning. In: Proceedings of the seventeenth international conference on machine learning. Morgan Kaufmann Publishers Inc., San Francisco, CA, USA, ICML ’00, pp 663–670

  28. You C, Lu J, Filev D et al (2019) Advanced planning for autonomous vehicles using reinforcement learning and deep inverse reinforcement learning. Robot Auton Syst 114:1–18. https://doi.org/10.1016/j.robot.2019.01.003

    Article  Google Scholar 

  29. Ziebart BD, Maas AL, Dey AK, et al (2008a) Navigate like a cabbie: probabilistic reasoning from observed context-aware behavior. In: Proceedings of the 10th international conference on ubiquitous computing. Association for Computing Machinery, New York, UbiComp ’08, pp 322–331

  30. Ziebart BD, Maas AL, Bagnell JA, et al (2008b) Maximum entropy inverse reinforcement learning. In: Proceedings of the 23rd AAAI conference on artificial intelligence. Association for the Advancement of Artificial Intelligence, New York, pp 1433–1438

  31. Morgenstern O, Von Neumann J (1953) Theory of games and economic behavior. Princeton University Press, Princeton

    Google Scholar 

  32. So W, Robbiano M, de Abreu NMM et al (2010) Applications of a theorem by Ky Fan in the theory of graph energy. Linear Algebra Appl 432(9):2163–2169. https://doi.org/10.1016/j.laa.2009.01.006

    Article  MathSciNet  Google Scholar 

  33. Liu CH, Dai Z, Zhao Y et al (2021) Distributed and energy-efficient mobile crowdsensing with charging stations by deep reinforcement learning. IEEE Trans Mob Comput 20(1):130–146. https://doi.org/10.1109/TMC.2019.2938509

    Article  Google Scholar 

  34. Vaswani A, Shazeer N, Parmar N, et al (2017) Attention is all you need. In: Proceedings of the 31st international conference on neural information processing systems. Curran Associates Inc., Red Hook, NIPS’17, pp 6000–6010

  35. Rossi RA, Ahmed NK (2015) The network data repository with interactive graph analytics and visualization. In: Bonet B, Koenig S (eds) Proceedings of the 29th AAAI conference on artificial intelligence, January 25-30, 2015, Austin, Texas. AAAI Press, pp 4292–4293

  36. Adler J, Lunz S (2018) Banach wasserstein gan. In: Proceedings of the 32nd international conference on neural information processing systems. Curran Associates Inc., Red Hook, NIPS’18, pp 6755–6764

  37. Junuthula RR, Xu KS, Devabhaktuni VK (2016) Evaluating link prediction accuracy in dynamic networks with added and removed edges. In: 2016 IEEE international conferences on big data and cloud computing (BDCloud), Social Computing and Networking (SocialCom), Sustainable Computing and Communications (SustainCom) (BDCloud-SocialCom-SustainCom), pp 377–384

  38. Pu C, Li J, Wang J et al (2022) The node-similarity distribution of complex networks and its applications in link prediction. IEEE Trans Knowl Data Eng 34(8):4011–4023

    Article  Google Scholar 

  39. Perozzi B, Al-Rfou R, Skiena S (2014) DeepWalk: online learning of social representations. In: Proceedings of the 20th ACM SIGKDD international conference on knowledge discovery and data mining. Association for Computing Machinery, New York, pp 701–710

Download references

Acknowledgements

This work was supported by two National Natural Science Foundation of China under Grant 61772102 and 62176036, and by the Science Foundation Ireland under grant SFI/12/RC/2289_P2.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Hongbo Liu.

Ethics declarations

Conflict of interest

All authors certify that they have no affiliations with or involvement in any organization or entity with any financial interest or non-financial interest in the subject matter or materials discussed in this manuscript.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Proof of Theorem

Proof of Theorem

The neighborhood structure vector maps the neighborhood information of each node included in the adjacency matrix to the vector space with fixed dimensions. In this section, we prove the injectivity of the mapping.

Theorem 1

(Injective mapping of node neighborhood structure) Given any one graph \(G=(V,E)\) and its adjacency matrix A, the neighborhood topological vector \(X_i=[x_{i1},x_{i2},\ldots ,x_{il}]\) of node \(v_i\) are obtained by summing the matrices \(\{A,A^2,\ldots ,A^l\}\) by row. If and only if the neighborhood structures of node \(v_i\) are isomorphic as the node \(v_j\), their neighborhood topological vectors are the same \(X_i=X_j\).

Proof

Firstly, the element \(x_{il}\) represents the number of different walks with length l from the node i. If \(l=2\), \(x_{i2}\) is equal to

$$\begin{aligned} \begin{aligned} x_{i2}&= \sum _{m=1}^{n}a_{im}^{(2)},\\ a_{im}^{(2)}&=\sum _{k=1}^{n}a_{ik}^{(1)}\times a_{km}^{(1)}, \end{aligned} \end{aligned}$$
(16)

where \(a_{ik}^{_{(1)}}\) and \(a_{im}^{_{(2)}}\) are the values of A[ij] and \(A^2[i,j]\), respectively. If only if \(a_{ik}^{_{(1)}}=1\) and \(a_{km}^{_{(1)}}=1\), \(a_{im}^{_{(2)}}=1\), namely, there exists a walk from the node \(v_i\) to node \(v_m\), \((i\rightarrow k\rightarrow m)\) in the graph G. Therefore, \(x_{i2}\) denotes the total number of walks with length 2 from node i to arbitrary nodes. Clearly, the value of \(X_i\) is solely determined by the neighborhood structure of node \(v_i\). Nodes with symmetrical structures are inherently guaranteed to possess identical neighborhood feature vectors.

Secondly, we emphasize the rationale behind the distinct neighborhood feature vectors corresponding to different node neighborhood structures. Two typical case of the distinct node neighborhood structures are shown in Fig. 11.

Fig. 11
figure 11

Two case of the difference on the neighborhood structures of nodes

Obviously in the Case 1, the discrepancy in the number of k-th hop neighboring nodes between the nodes 1 and a results in distinct values for \(x_{1k}\) and \(x_{ak}\). Therefore, we need to focus on the Case 2 where the number of k-th hop neighbors is the same for two nodes, but the neighborhood structures are different.

To formalize the neighborhood structure vector, we assume \(d_0\), \(d_1\) and \(d_2\) represent the root node \(v_i\), its neighbor nodes and its 2-hop nodes, respectively. Then, the value of \(x_{i4}\) in the neighborhood topological vector \(X_i\) is

$$\begin{aligned} x_4^i=\left\{ \begin{array}{lr} (d_0\rightarrow d_1\rightarrow d_2\rightarrow d_1\rightarrow d_2)\\ (d_0\rightarrow d_1\rightarrow d_0\rightarrow d_1\rightarrow d_2)\\ (d_0\rightarrow d_1\rightarrow d_0\rightarrow d_1\rightarrow d_0)\\ (d_0\rightarrow d_1\rightarrow d_2\rightarrow d_1\rightarrow d_0) \end{array} \right\} , \end{aligned}$$
(17)

where \((d_0\rightarrow d_1\rightarrow d_2\rightarrow d_1\rightarrow d_2)\) means all the walks from the root node \(v_i\) to the 2-hop nodes when the length of walks is equal to 4. We can observe that the number of walks \((d_0\rightarrow d_1\rightarrow d_2\rightarrow d_1\rightarrow d_2)\) is related to the allocation of 2-hop neighbor nodes. The reason of \(x_{14}\ne x_{a4}\) is

$$\begin{aligned} \left\{ \begin{array}{lr} (1\rightarrow 5\rightarrow 7\rightarrow 5\rightarrow 7)\\ (1\rightarrow 6\rightarrow 8\rightarrow 6\rightarrow 8) \end{array} \right\} \ne \left\{ \begin{array}{lr} (a\rightarrow f\rightarrow g\rightarrow f\rightarrow g)\\ (a\rightarrow f\rightarrow g\rightarrow f\rightarrow h)\\ (a\rightarrow f\rightarrow h\rightarrow f\rightarrow h)\\ (a\rightarrow f\rightarrow h\rightarrow f\rightarrow g) \end{array} \right\} , \end{aligned}$$
(18)

Furthermore, we suppose \(|d_1|=n,|d_2|=m\) represent the number of neighbor nodes and 2-hop nodes of the node \(v_i\), respectively. And the allocation of 2-hop nodes is

$$\begin{aligned} m=k_1+k_2+\dots +k_n,k_s\in \{0,1,\ldots ,m\}, \end{aligned}$$
(19)

where \(k_s\) represents the number of 2-hop nodes connected with the s-th neighbor node. Then, Eq. (17) is changed into

$$\begin{aligned} x_{i4}=\sum _{s=1}^{n}k_s^2+nm+n^2+m, \end{aligned}$$
(20)

where \(\sum _{s=1}^{n}k_s^2\) is the number of walks \((d_0\rightarrow d_1\rightarrow d_2\rightarrow d_1\rightarrow d_2)\), in which both \(d_1^{_{(s)}}\rightarrow d_2\) and \(d_2\rightarrow d_1^{_{(s)}}\rightarrow d_2\) have \(k_s\) walks, \(d_1^{_{(s)}}\in d_1\). For example, \(x_{a4}=(2^2+0^2+2^2)+(3\times 4+3^2+4)=33\), \(x_{14}=(2^2+1^2+1^2)+(3\times 4+3^2+4)=31\). From Eq. (20), when the nodes \(v_i\) and \(v_j\) have the same number of 2-hop neighbors but different allocation methods, the necessary condition of \(x_{i4}=x_{j4}\) is,

$$\begin{aligned} \begin{aligned} k_{i1}+k_{i2}+\ldots +k_{in}&=k_{j1}+k_{j2}+\ldots +k_{jn}=m,\\ k_{i1}^2+k_{i2}^2+\ldots +k_{in}^2&= k_{j1}^2+k_{j2}^2+\ldots +k_{jn}^2. \end{aligned} \end{aligned}$$
(21)

However, Eq. (21) does have a positive integer solution, for example, if \(m=6\), \(\{k_{i1},k_{i2},k_{i3},k_{i4}\}=\{2,2,2,0\}\) and \(\{k_{j1},k_{j2},k_{j3},k_{j4}\}=\{3,1,1,1\}\). Thus we must consider setting a larger l, such as the value of \(x_{i6}\) is

$$\begin{aligned} x_{i6}=\sum _{s=1}^{n}k_s^3+\sum _{s=1}^{n}k_s^2+O(n,m), \end{aligned}$$
(22)

where O(nm) represents the polynomial containing parameters n and m. The reason for the occurrence of the third power of \(\{k_s\}\) in Eq. (22) is due to the special walk shown in Fig. 12. Simply put, there are \(k_s\) possibilities each time a random walk moves from the s-th neighboring node to its 2-hop nodes.

Fig. 12
figure 12

The special walk with \((d_0,\overline{d_1,d_2},\ldots )\)

Then, when \(\{k_{i1},k_{i2},\ldots ,k_{in}\}\ne \{k_{j1},k_{j2},\ldots ,k_{jn}\}\), \(l=2N\), \(|d_1|=n\), \(|k_i|=|k_j|=m\) and \(X_i=X_j\), the following equations must be met:

$$\begin{aligned} {\left\{ \begin{array}{ll} \sum _{s=1}^{n}(k_{is})^2=\sum _{s=1}^{n}(k_{js})^2,\\ \sum _{s=1}^{n}(k_{is})^3=\sum _{s=1}^{n}(k_{js})^3,\\ \hspace{5mm}\vdots \hspace{22mm}\vdots \\ \sum _{s=1}^{n}(k_{is})^N=\sum _{s=1}^{n}(k_{js})^N, \end{array}\right. } \end{aligned}$$
(23)

where \(k_{is},k_{js}\subset \mathbf {N^*}\), and Eq. (23) can be expressed as an integer programming problem. More importantly, n and m are are two specific positive integers, both of which are smaller than the total number of nodes in the entire network (typically in the range of hundreds to thousands). Therefore, according to the combination theory, the number of solutions in the feasible solution space is finite, and it is

$$\begin{aligned} \Omega (k_i,k_j)=\left( {\begin{array}{c}\left( {\begin{array}{c}m\\ n\end{array}}\right) \\ 2\end{array}}\right) . \end{aligned}$$
(24)

Furthermore, as the value of l increases, it indicates a higher number of constraints in the integer problem of Eq. (23), resulting in a lower probability of finding feasible solutions. In other words, there is a injective mapping from the neighborhood structure to the neighborhood topological vector \(\{x_{i1},x_{i2},\ldots ,x_{in}\}\), when l is set as a sufficiently large integer. \(\square\)

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Jiang, X., Liu, H., Yang, L. et al. Unraveling human social behavior motivations via inverse reinforcement learning-based link prediction. Computing 106, 1963–1986 (2024). https://doi.org/10.1007/s00607-024-01279-w

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00607-024-01279-w

Keywords

Mathematics Subject Classification