Skip to main content
Log in

Graph autoencoder for directed weighted network

  • Data analytics and machine learning
  • Published:
Soft Computing Aims and scope Submit manuscript

Abstract

Network embedding technology transforms network structure into node vectors, which reduces the complexity of representation and can be effectively applied to tasks such as classification, network reconstruction and link prediction. The main concern of network embedding is to keep the local structural features while effectively capturing the global features of the network. The “shallow” network representation models cannot capture the deep nonlinear features of the network, and the generated network embedding is usually not the optimal solution. In this paper, a new graph autoencoder-based network representation model combines the first- and second-order proximity to evaluate the performance of network embedding. Aiming at the shortcomings of existing network representation methods in weighted and directed networks, on one hand, the concepts of receiving vector and sending vector are introduced with a simplification of decoding part of the neural network which reduces computation complexity; on the other hand, a measurement index based on node degree is proposed to better emphasize the weighted information in the application of network representation. Experiments including directed weighted networks and undirected unweighted networks show that the proposed method achieves better results than the baseline methods for network reconstruction and link prediction tasks and is of higher computation efficiency than previous graph autoencoder algorithms. Besides, the proposed weighted index is able to improve performances of all baseline methods as an external assistance.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11

Similar content being viewed by others

References

  • Abu-El-Haija S, Perozzi B, Al-Rfou R (2017) Learning edge representations via low-rank asymmetric projections. In: Proceedings of the 2017 ACM on conference on information and knowledge management. ACM, pp 1787–1796

  • Barabasi A-L, Albert R (1999) Emergence of scaling in random networks. Science 286(5439):509–512

  • Belkin M (2001) Laplacian eigenmaps and spectral techniques for embedding and clustering. In: Nips, vol 14, pp 585–591

  • Cunchao T, Cheng Y, Zhiyuan L et al (2017) Network representation learning: an overview. Sci Sin (Informationis) 47:980–996 ((in Chinese))

    Google Scholar 

  • Goldberg Y (2016) A primer on neural network models for natural language processing. J Artif Intell Res 57:345–420

  • Goyal P, Ferrara E (2018) graph embedding techniques, applications, and performance: a survey. Knowl Based Syst 151:78–94

  • Goyal P, Kamra N, He X, Liu Y (2018) Dyngem: deep embedding method for dynamic graphs. arXiv:1805.11273

  • Grover A, Leskovec J (2016) node2vec: scalable feature learning for networks. In: Proceedings of the 22nd ACM SIGKDD international conference on Knowledge discovery and data mining. ACM, pp 855–864

  • Hamilton WL, Ying R, Leskovec J (2017) Representation learning on graphs: methods and applications. arXiv:1709.05584

  • Japkowicz N, Shah M (2011) Evaluating learning algorithms: a classification perspective. Cambridge University Press, Cambridge

    Book  Google Scholar 

  • Khosla M, Anand A, Setty V (2019) A comprehensive comparison of unsupervised network representation learning methods. arXiv:1903.07902

  • Kipf TN, Welling M (2016) Variational graph auto-encoders. . arXiv:1611.07308

  • Klimt B, Yang Y (2004) The enron corpus: a new dataset for email classification research. Ecml 3201:217–226

    MATH  Google Scholar 

  • Lin-Yuan L (2010) Link Prediction on Complex Networks. J Univ Electron Sci Technol China 39(5):651–661

  • Linyuan L, Tao Z (2010) Link prediction in complex networks: a survey. Physica A Stat Mech Appl 390(6):1150–1170

    Google Scholar 

  • Perozzi B, Al-Rfou R, Skiena S (2014) Deepwalk: online learning of social representations. In: Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining. ACM, pp 701–710

  • Salha G, Limnios S, Hennequin R, et al (2019) Gravity-inspired graph autoencoders for directed link prediction. In: Proceedings of the 28th ACM international conference on information and knowledge management, pp 589–598

  • Sutskever I, Vinyals O, Le QV (2014) Sequence to sequence learning with neural networks. Adv Neural Inform Process Syst 3104–3112

  • Tan Q, Liu N, Hu X (2019) Deep representation learning for social network analysis. Front Big Data 2

  • Wang D, Peng C, Zhu W (2016) structural deep network embedding. In: Acm Sigkdd international conference on knowledge discovery and data mining

  • Yang M, Xingxing L, Jincai H, et al (2017) Intercity transportation construction based on link prediction. In: 2017 IEEE 29th international conference on tools with artificial intelligence

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Guangquan Cheng.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Ma, Y., Li, Y., Liang, X. et al. Graph autoencoder for directed weighted network. Soft Comput 26, 1217–1230 (2022). https://doi.org/10.1007/s00500-021-06580-w

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00500-021-06580-w

Keywords

Navigation