Skip to main content

Advertisement

GraphSense: a self-aware dynamic graph learning networks for graph data over internet

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

Dynamic graph data learning is an important data analysis technique. In the age of big data, the volume of data produced daily is immense, the data types are varied, the value density is low, and the data continues to accumulate over time. These characteristics make data processing more challenging. In particular, unstructured data, unlike structured data, does not have a fixed format, and its volume is large and variable, which presents a significant challenge to traditional data processing techniques. Nowadays, researchers have been employing graph neural network models to analyze unstructured data. However, real-world graph structures are dynamic and time-varying, and the static graph neural network cannot effectively learn graph node embeddings and network structures. To address the challenges mentioned above, we propose a self-aware dynamic graph network structure learning model, called GraphSense. The algorithm consists of two modules: self-sensing neighborhood aggregation algorithm and dynamic graph structure learning algorithm based on RNN. GraphSense can make each node discover more valuable neighbors through the self-aware neighborhood aggregation algorithm in each epoch. The algorithm employs gated recurrent unit to dynamically aggregate the information of node neighbors to learn the high-order information. Next, in order to capture the temporal properties of graph structures, we employ dynamic graph structure learning algorithm based on RNN to replicate the time evolution process of dynamic graphs. Finally, we evaluate the performance of GraphSense on four publicly available datasets by two specific tasks(edge and node classification). The experimental results show that the proposed GraphSense model outperforms the baseline model by 2.0% to 25.0% on the Elliptic dataset, 2.5% to 27.0% on the Bitcoin-alpha dataset, 3.0% to 31.0% on the Bitcoin-otc dataset, and 0.9% to 26.0% on the Reddit dataset in terms of F1 scores. The results suggest that our model is effective in learning from dynamic graph data.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Algorithm 1
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

Data Availability

Not Applicable.

References

  1. Wan C, Jones DT (2020) Protein function prediction is improved by creating synthetic feature samples with generative adversarial networks. Nat Mach Intell. 2(9):540–550

    Article  MATH  Google Scholar 

  2. Zhang J, Liu Y, Ding G, Tang B, Chen Y (2024) Adaptive decomposition and extraction network of individual fingerprint features for specific emitter identification. IEEE Trans Inf Foren Secur

  3. Zhang J, Su Q, Tang B, Wang C, Li Y (2021) Dpsnet: Multitask learning using geometry reasoning for scene depth and semantics. IEEE Trans Neural Netw Learn Syst. 34(6):2710–2721

    Article  MATH  Google Scholar 

  4. Li X, Chen M, Hu Y, Zhang Z, Liu D, Mao S (2024) Jointly optimizing terahertz based sensing and communications in vehicular networks: A dynamic graph neural network approach. IEEE Trans Wireless Commu

  5. Duan M, Zheng T, Gao Y, Wang G, Feng Z, Wang X (2024) Dga-gnn: Dynamic grouping aggregation gnn for fraud detection. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol 38, pp 11820–11828

  6. Lu R, Cheng Z, Chen B, Yuan X (2024) Motion-aware dynamic graph neural network for video compressive sensing. IEEE Trans Pattern Anal Mach Intell

  7. Jafarian B, Yazdani N, Haghighi MS (2024) Using attentive temporal gnn for dynamic trust assessment in the presence of malicious entities. Exp Syst Appl, 125391

  8. Zhou B, Zhou H, Wang W, Chen L, Ma J, Zheng Z (2024) Hdm-gnn: A heterogeneous dynamic multi-view graph neural network for crime prediction. ACM Trans Sensor Netw

  9. Ying Z, Bourgeois D, You J, Zitnik M, Leskovec J (2019) Gnnexplainer: Generating explanations for graph neural networks. Adv Neural Inf Process Syst. 32

  10. Yin N, Wang M, Chen Z, De Masi G, Xiong H, Gu, B (2024) Dynamic spiking graph neural networks. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol 38, pp 16495–16503

  11. Liu J, Liu J, Zhao K, Tang Y, Chen W (2024) Tp-gnn: Continuous dynamic graph neural network for graph classification. In: 2024 IEEE 40th International Conference on Data Engineering (ICDE), pp 2848–2861 . IEEE

  12. Yang S, Hu B, Zhang Z, Sun W, Wang Y, Zhou J, Shan H, Cao Y, Ye B, Fang Y, et al. (2021) Inductive link prediction with interactive structure learning on attributed graph. In: Machine Learning and Knowledge Discovery in Databases. Research Track: European Conference, ECML PKDD 2021, Bilbao, Spain, 13–17, September 2021, Proceedings, Part II 21, pp 383–398 . Springer

  13. Kipf TN, Welling M (2016) Semi-supervised classification with graph convolutional networks

  14. Skarding J, Gabrys B, Musial K (2021) Foundations and modeling of dynamic networks using dynamic graph neural networks: A survey. IEEE Access. 9, 79143–79168. https://doi.org/10.1109/ACCESS.2021.3082932

  15. Ma Y, Guo Z, Ren Z, Tang J, Yin D (2020) Streaming graph neural networks. In: Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, pp 719–728

  16. Pareja A, Domeniconi G, Chen J, Ma T, Suzumura T, Kanezashi H, Kaler T, Schardl T, Leiserson C (2020) Evolvegcn: Evolving graph convolutional networks for dynamic graphs. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol 34, pp 5363–5370

  17. Bahdanau D, Cho K, Bengio Y (2014) Neural machine translation by jointly learning to align and translate. arXiv:1409.0473

  18. Xu K, Ba J, Kiros R, Cho K, Courville A, Salakhudinov R, Zemel R, Bengio Y (2015) Show, attend and tell: Neural image caption generation with visual attention. In: International Conference on Machine Learning, pp 2048–2057 . PMLR

  19. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I (2017) Attention is all you need. Adv Neural Inf Process Syst. 30

  20. Liu Y, Lapata M (2018) Learning structured text representations. Trans Ass Comput Linguist. 6:63–75

    MATH  Google Scholar 

  21. Ying C, Cai T, Luo S, Zheng S, Ke G, He D, Shen Y, Liu T-Y (2021) Do transformers really perform badly for graph representation? Adv Neural Inf Process Syst. 34:28877–28888

    MATH  Google Scholar 

  22. Sankar A, Wu Y, Gou L, Zhang W, Yang H (2020) Dysat: Deep neural representation learning on dynamic graphs via self-attention networks. In: Proceedings of the 13th International Conference on Web Search and Data Mining, pp 519–527

  23. Perozzi B, Al-Rfou R, Skiena S (2014) Deepwalk: Online learning of social representations. In: Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp 701–710

  24. Grover A, Leskovec J (2016) node2vec: Scalable feature learning for networks. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp 855–864

  25. Narayanan A, Chandramohan M, Venkatesan R, Chen L, Liu Y, Jaiswal S (2017) graph2vec: Learning distributed representations of graphs. arXiv:1707.05005

  26. Tang J, Qu M, Wang M, Zhang M, Yan J, Mei Q (2015) Line: Large-scale information network embedding. In: Proceedings of the 24th International Conference on World Wide Web, pp 1067–1077

  27. Hamilton W, Ying Z, Leskovec J (2017) Inductive representation learning on large graphs. Adv Neural Inf Process Syst. 30

  28. Gao H, Wang Z, Ji S (2018) Large-scale learnable graph convolutional networks. In: Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pp 1416–1424

  29. Veličković P, Cucurull G, Casanova A, Romero A, Lio P, Bengio Y (2017) Graph attention networks. arXiv:1710.10903

  30. Lee J, Lee I, Kang J (2019) Self-attention graph pooling. In: International Conference on Machine Learning, pp 3734–3743 . PMLR

  31. Nguyen DQ, Nguyen TD, Phung D (2022) Universal graph transformer self-attention networks. In: Companion Proceedings of the Web Conference 2022, pp 193–196

  32. Yu W, Aggarwal CC, Wang W (2017) Temporally factorized network modeling for evolutionary network analysis. In: Proceedings of the Tenth ACM International Conference on Web Search and Data Mining, pp 455–464

  33. Seo Y, Defferrard M, Vandergheynst P, Bresson X (2018) Structured sequence modeling with graph convolutional recurrent networks. In: Neural Information Processing: 25th International Conference, ICONIP 2018, Siem Reap, Cambodia, 13-16 December 2018, Proceedings, Part I 25, pp 362–373 . Springer

  34. Kumar S, Zhang X, Leskovec J (2019) Predicting dynamic embedding trajectory in temporal interaction networks. In: Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pp 1269–1278

  35. Chen J, Wang X, Xu X (2022) Gc-lstm: Graph convolution embedded lstm for dynamic network link prediction. Applied Intell, 1–16

  36. Bonner S, Atapour-Abarghouei A, Jackson PT, Brennan J, Kureshi I, Theodoropoulos G, McGough AS, Obara B (2019) Temporal neighbourhood aggregation: Predicting future links in temporal graphs via recurrent variational graph convolutions. In: 2019 IEEE International Conference on Big Data (Big Data), pp 5336–5345 . IEEE

  37. Taheri A, Gimpel K, Berger-Wolf T (2019) Learning to represent the evolution of dynamic graphs with recurrent models. In: Companion Proceedings of the 2019 World Wide Web Conference, pp 301–307

  38. Gao C, Zhu J, Zhang F, Wang Z, Li X (2022) A novel representation learning for dynamic graphs based on graph convolutional networks. IEEE Trans Cyber

  39. Zhang J, Zhang H, Xia C, Sun L (2020) Graph-bert: Only attention is needed for learning graph representations. arXiv:2001.05140

  40. Wu J, Cao M, Cheung JCK, Hamilton WL (2020) Temp: Temporal message passing for temporal knowledge graph completion. arXiv:2010.03526

  41. Xu D, Ruan C, Korpeoglu E, Kumar S, Achan K (2020) Inductive representation learning on temporal graphs. arXiv:2002.07962

  42. Li H, Li C, Feng K, Yuan Y, Wang G, Zha H (2024) Robust knowledge adaptation for dynamic graph neural networks. IEEE Trans Knowl Data Eng

  43. Kumar S, Spezzano F, Subrahmanian V, Faloutsos C (2016) Edge weight prediction in weighted signed networks. In: 2016 IEEE 16th International Conference on Data Mining (ICDM), pp 221–230 . IEEE

  44. Kumar S, Hooi B, Makhija D, Kumar M, Faloutsos C, Subrahmanian V (2018) Rev2: Fraudulent user prediction in rating platforms. In: Proceedings of the Eleventh ACM International Conference on Web Search and Data Mining, pp 333–341

  45. Weber M, Domeniconi G, Chen J, Weidele DKI, Bellei C, Robinson T, Leiserson CE (2019) Anti-money laundering in bitcoin: Experimenting with graph convolutional networks for financial forensics. arXiv:1908.02591

  46. Kumar S, Hamilton WL, Leskovec J, Jurafsky D (2018) Community interaction and conflict on the web. In: Proceedings of the 2018 World Wide Web Conference, pp 933–943

  47. Zhao X, Chen F, Cho J-H (2018) Deep learning for predicting dynamic uncertain opinions in network data. In: 2018 IEEE International Conference on Big Data (Big Data), pp 1150–1155 . IEEE

  48. Chengqing Y, Guangxi Y, Chengming Y, Yu Z, Xiwei M (2023) A multi-factor driven spatiotemporal wind power prediction model based on ensemble deep graph attention reinforcement learning networks. Energy. 263:126034

  49. Chen C, Geng H, Yang N, Yang X, Yan J (2024) Easydgl: Encode, train and interpret for continuous-time dynamic graph learning. IEEE Trans Pattern Anal Mach Intell

Download references

Funding

This research was funded by the National Key Research and Development Program Project grant number 2020YFB1005503, the Natural Science Foundation of Jiangsu Province grant number BK20201415.

Author information

Authors and Affiliations

Authors

Contributions

Zhiyuan Li gave the idea of this paper and model design, and wrote the main manuscript text. En-Han He and Ying-Yi Zhou finished the simulation experiment validation and experimental writtings. All authors have reviewed and agreed to the published version of the manuscript.

Corresponding author

Correspondence to Zhi-Yuan Li.

Ethics declarations

Conflicts of Interest

The authors declare no conflict of interest.

Ethics Approval

Not applicable.

Consent for publication

We agree to publish on the Journal of Peer-to-Peer Networking and Applications.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Li, ZY., Zhou, YY. & He, EH. GraphSense: a self-aware dynamic graph learning networks for graph data over internet. Appl Intell 55, 41 (2025). https://doi.org/10.1007/s10489-024-05882-4

Download citation

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s10489-024-05882-4

Keywords