Skip to main content

Alleviating Over-Smoothing via Aggregation over Compact Manifolds

  • Conference paper
  • First Online:
Advances in Knowledge Discovery and Data Mining (PAKDD 2024)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 14646))

Included in the following conference series:

  • 194 Accesses

Abstract

Graph neural networks (GNNs) have achieved significant success in various applications. Most GNNs learn the node features with information aggregation of its neighbors and feature transformation in each layer. However, the node features become indistinguishable after many layers, leading to performance deterioration: a significant limitation known as over-smoothing. Past work adopted various techniques for addressing this issue, such as normalization and skip-connection of layer-wise output. After the study, we found that the information aggregations in existing work are all contracted aggregations, with the intrinsic property that features will inevitably converge to the same single point after many layers. To this end, we propose the aggregation over compacted manifolds method (ACM) that replaces the existing information aggregation with aggregation over compact manifolds, a special type of manifold, which avoids contracted aggregations. In this work, we theoretically analyze contracted aggregation and its properties. We also provide an extensive empirical evaluation that shows ACM can effectively alleviate over-smoothing and outperforms the state-of-the-art. The code can be found in https://github.com/DongzhuoranZhou/ACM.git.

D. Zhou and H. Yang—Equal contribution.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 119.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 79.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    In Fig. 1b, the unit circle is used as the embedding space. The aggregation of points is defined under polar coordinates. For example, the aggregation of two points in the unit circle with polar coordinate \((1, \theta ), (1, \phi )\) is \((1, \frac{\theta +\phi }{2})\).

References

  1. Bachmann, G., Bécigneul, G., Ganea, O.: Constant curvature graph convolutional networks. In: ICML, pp. 486–496. PMLR (2020)

    Google Scholar 

  2. Balazevic, I., Allen, C., Hospedales, T.M.: Multi-relational poincaré graph embeddings. In: NeurIPS, pp. 4465–4475 (2019)

    Google Scholar 

  3. Cai, C., Wang, Y.: A note on over-smoothing for graph neural networks. CoRR abs/ arXiv: 2006.13318 (2020)

  4. Chami, I., et al: Hyperbolic graph convolutional neural networks. In: NeurIPS, pp. 4869–4880 (2019)

    Google Scholar 

  5. Chen, D., Lin, Y., et al.: Measuring and relieving the over-smoothing problem for graph neural networks from the topological view. CoRR abs/ arXiv: 1909.03211 (2019)

  6. Chen, M., Wei, Z., Huang, Z., Ding, B., Li, Y.: Simple and deep graph convolutional networks. In: ICML, pp. 1725–1735. PMLR (2020)

    Google Scholar 

  7. Chien, E., Peng, J., Li, P., Milenkovic, O.: Adaptive universal generalized pagerank graph neural network. In: ICLR (2021)

    Google Scholar 

  8. Ganea, O., Bécigneul, G., Hofmann, T.: Hyperbolic neural networks. In: NeurIPS, pp. 5350–5360 (2018)

    Google Scholar 

  9. Gao, H., Wang, Z., Ji, S.: Large-scale learnable graph convolutional networks. In: KDD, pp. 1416–1424. ACM (2018)

    Google Scholar 

  10. Glorot, X., Bengio, Y.: Understanding the difficulty of training deep feedforward neural networks. In: AISTATS, pp. 249–256 (2010)

    Google Scholar 

  11. Gülçehre, Ç., et al.: Hyperbolic attention networks. In: ICLR (2019)

    Google Scholar 

  12. Hamilton, W.L.: Graph Representation Learning. Synthesis Lect. Artifi. Intell. Mach. Learn. (2020)

    Google Scholar 

  13. Hamilton, W.L., et al.: Inductive representation learning on large graphs. In: NIPS, pp. 1024–1034 (2017)

    Google Scholar 

  14. He, K., et al.: Deep residual learning for image recognition. In: CVPR, pp. 770–778. IEEE Computer Society (2016)

    Google Scholar 

  15. Hou, Y., Zhang, J., et al.: Measuring and improving the use of graph information in graph neural networks. In: ICLR (2020)

    Google Scholar 

  16. Huang, W., et al.: Tackling over-smoothing for general graph convolutional networks. CoRR (2020)

    Google Scholar 

  17. Ioffe, S., Szegedy, C.: Batch normalization: accelerating deep network training by reducing internal covariate shift. In: ICML, pp. 448–456 (2015)

    Google Scholar 

  18. Jin, W., Et al.: Feature overcorrelation in deep graph neural networks: a new perspective. In: KDD, pp. 709–719. ACM (2022)

    Google Scholar 

  19. Khrulkov, V., Et al.: Hyperbolic image embeddings. In: CVPR, pp. 6417–6427. Computer Vision Foundation/IEEE (2020)

    Google Scholar 

  20. Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. In: ICLR (2015)

    Google Scholar 

  21. Kipf, T.N., Welling, M.: Semi-supervised classification with graph convolutional networks. CoRR abs/ arxiv: 1609.02907 (2016)

  22. Klicpera, J., Et al.: Predict then propagate: graph neural networks meet personalized pagerank. In: ICLR (2019)

    Google Scholar 

  23. . Klicpera, J., et al.: Predict then propagate: graph neural networks meet personalized pagerank. In: ICLR (2019)

    Google Scholar 

  24. Lee, J.: Introduction to Smooth Manifolds. Graduate Texts in Mathematics

    Google Scholar 

  25. Li, G., Müller, M., et al.: Deepgcns: can gcns go as deep as cnns? In: ICCV, pp. 9266–9275. IEEE (2019)

    Google Scholar 

  26. Li, Q., Han, Z., Wu, X.M.: Deeper insights into graph convolutional networks for semi-supervised learning. In: AAAI, pp. 3538–3545 (2018)

    Google Scholar 

  27. Liu, M., Gao, H., Ji, S.: Towards deeper graph neural networks. In: KDD, pp. 338–348. ACM (2020)

    Google Scholar 

  28. Mendelson, B.: Introduction to topology (1990)

    Google Scholar 

  29. Oono, K., Suzuki, T.: Graph neural networks exponentially lose expressive power for node classification. In: ICLR (2020)

    Google Scholar 

  30. Pei, H., et al.: Geom-gcn: Geometric graph convolutional networks. In: ICLR (2020)

    Google Scholar 

  31. Rashid, A.M., Karypis, G., et al.: Learning preferences of new users in recommender systems: an information theoretic approach. SIGKDD Explor., 90–100 (2008)

    Google Scholar 

  32. Rong, Y., Huang, W., Xu, T., Huang, J.: Dropedge: towards deep graph convolutional networks on node classification. In: ICLR (2020)

    Google Scholar 

  33. Shchur, O., Mumme, M., Bojchevski, A., Günnemann, S.: Pitfalls of graph neural network evaluation. CoRR abs/ arXiv: 1811.05868 (2018)

  34. Srivastava, N., et al.: Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15(1), 1929–1958 (2014)

    MathSciNet  Google Scholar 

  35. Ungar, A.A.: Barycentric calculus in Euclidean and hyperbolic geometry: a comparative introduction (2010)

    Google Scholar 

  36. Velickovic, P., Cucurull, G., Casanova, A., Romero, A., Liò, P., Bengio, Y.: Graph attention networks. CoRR abs/ arXiv: 1710.10903 (2017)

  37. Wu, F., et al.: Simplifying graph convolutional networks. In: ICML, vol. 97, pp. 6861–6871. PMLR (2019)

    Google Scholar 

  38. Xu, K., Hu, W., et al.: How powerful are graph neural networks? In: ICLR (2019)

    Google Scholar 

  39. Xu, K., Li, C., Tian, Y., et al.: Representation learning on graphs with jumping knowledge networks. In: ICML, pp. 5449–5458. PMLR (2018)

    Google Scholar 

  40. Yang, Z., et al.: Revisiting semi-supervised learning with graph embeddings. In: ICML. JMLR Workshop and Conference Proceedings, vol. 48, pp. 40–48 (2016)

    Google Scholar 

  41. Zhao, L., Akoglu, L.: Pairnorm: tackling oversmoothing in gnns. In: ICLR (2020)

    Google Scholar 

  42. Zhou, J., Cui, G., et al.: Graph neural networks: a review of methods and applications. AI Open 1, 57–81 (2020)

    Article  Google Scholar 

  43. Zhou, K., et al.: Towards deeper graph neural networks with differentiable group normalization. In: NeurIPS (2020)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Dongzhuoran Zhou .

Editor information

Editors and Affiliations

1 Electronic supplementary material

Below is the link to the electronic supplementary material.

Supplementary material 1 (pdf 1350 KB)

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Zhou, D., Yang, H., Xiong, B., Ma, Y., Kharlamov, E. (2024). Alleviating Over-Smoothing via Aggregation over Compact Manifolds. In: Yang, DN., Xie, X., Tseng, V.S., Pei, J., Huang, JW., Lin, J.CW. (eds) Advances in Knowledge Discovery and Data Mining. PAKDD 2024. Lecture Notes in Computer Science(), vol 14646. Springer, Singapore. https://doi.org/10.1007/978-981-97-2253-2_31

Download citation

  • DOI: https://doi.org/10.1007/978-981-97-2253-2_31

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-97-2252-5

  • Online ISBN: 978-981-97-2253-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics