Skip to main content
Log in

SSKGE: a time-saving knowledge graph embedding framework based on structure enhancement and semantic guidance

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

In knowledge graph embedding, an attempt is made to embed the objective facts and relationships expressed in the form of triplets into multidimensional vector space, facilitating various applications, such as link prediction and question answering. Structure embedding models focus on the graph structure while the importance of language semantics in inferring similar entities and relations is ignored. Semantic embedding models use pretrained language models to learn entity and relation embeddings based on text information, but they do not fully exploit graph structures that reflect relation patterns and mapping attributes. Structure and semantic information in knowledge graphs represent different hierarchical properties that are indispensable for comprehensive knowledge representation. In this paper, we propose a general knowledge graph embedding framework named SSKGE, which considers both the graph structure and language semantics and learns these two complementary characteristics to integrate entity and relation representations. To compensate for semantic embedding approaches that ignore the graph structure, we first design a structure loss function to explicitly model the graph structure attributes. Second, we leverage a pretrained language model that has been fine-tuned by the structure loss to guide the structure embedding approaches in enhancing the semantic information they lack and obtaining universal knowledge representations. Specifically, guidance is provided by a distance function that makes the spatial distribution of the two types of graph embeddings have a certain similarity. SSKGE significantly reduces the time cost of using a pretrained language model to complete a knowledge graph. Common knowledge graph embedding models such as TransE, DistMult, ComplEx, RotatE, PairRE, and HousE have achieved better results with multiple datasets, including FB15k, FB15k-237, WN18, and WN18RR, using the SSKGE framework. Extensive experiments and analyses have verified the effectiveness and practicality of SSKGE.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5

Similar content being viewed by others

Data Availibility Statement

The datasets generated and/or analysed during the current study will be made available on reasonable request.

Notes

  1. https://huggingface.co/bert-base-cased

References

  1. Fang Q, Zhang X, Hu J, Wu X, Xu C (2022) Contrastive multi-modal knowledge graph representation learning. IEEE Transactions on Knowledge and Data Engineering

  2. Ji S, Pan S, Cambria E, Marttinen P, Philip S (2021) A survey on knowledge graphs: Representation, acquisition, and applications. IEEE Trans Neural Netw Learn Syst 33(2):494–514

    Article  MathSciNet  Google Scholar 

  3. Rossi A, Barbosa D, Firmani D, Matinata A, Merialdo P (2021) Knowledge graph embedding for link prediction: A comparative analysis. ACM Trans Knowl Disc Data (TKDD) 15(2):1–49

    Article  Google Scholar 

  4. Tiddi I, Schlobach S (2022) Knowledge graphs as tools for explainable machine learning: A survey. Art Intell 302:103627

    Article  MathSciNet  MATH  Google Scholar 

  5. Bordes A, Usunier N, Garcia-Duran A, Weston J, Yakhnenko O (2013) Translating embeddings for modeling multi-relational data. Adv Neural Inf Process Syst 26

  6. Sun Z, Deng Z-H, Nie J-Y, Tang J (2019) Rotate: Knowledge graph embedding by relational rotation in complex space. arXiv:1902.10197

  7. Chao L, He J, Wang T, Chu W (2021) Pairre: Knowledge graph embeddings via paired relation vectors. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 4360–4369

  8. Nickel M, Tresp V, Kriegel H-P (2011) A three-way model for collective learning on multi-relational data. In Proceedings of the 28th International Conference on International Conference on Machine Learning, pages 809–816

  9. Yang B, Yih W-T, He X, Gao J, Deng L (2014) Embedding entities and relations for learning and inference in knowledge bases. arXiv:1412.6575

  10. Trouillon T, Welbl J, Riedel S, Éric Gaussier, Bouchard G (2016) Complex embeddings for simple link prediction. In International conference on machine learning, pages 2071–2080. PMLR

  11. Dettmers T, Minervini P, Stenetorp P, Riedel S (2018) Convolutional 2d knowledge graph embeddings. In Proceedings of the AAAI conference on artificial intelligence, volume 32

  12. Schlichtkrull M, Kipf T N, Bloem P, van den Berg R, Titov I, Welling M (2018) Modeling relational data with graph convolutional networks. In European semantic web conference, pages 593–607. Springer

  13. Shang C, Tang Y, Huang J, Bi J, He X, Zhou B (2019) End-to-end structure-aware convolutional networks for knowledge base completion. In Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence and Thirty-First Innovative Applications of Artificial Intelligence Conference and Ninth AAAI Symposium on Educational Advances in Artificial Intelligence, pages 3060–3067

  14. Wang B, Shen T, Long G, Zhou T, Wang Y, Chang Y (2021) Structure-augmented text representation learning for efficient knowledge graph completion. In Proceedings of the Web Conference 2021:1737–1748

    Google Scholar 

  15. Wang X, Gao T, Zhu Z, Zhang Z, Liu Z, Li J, Tang J (2021) Kepler: A unified model for knowledge embedding and pre-trained language representation. Trans Assoc Comput Linguistics 9:176–194

  16. Liu X, Hussain H, Razouk H, Kern R (2022) Effective use of bert in graph embeddings for sparse knowledge graph completion. In Proceedings of the 37th ACM/SIGAPP Symposium on Applied Computing, pages 799–802

  17. Yao L, Mao C, Luo Y (2019) Kg-bert: Bert for knowledge graph completion. arXiv:1909.03193

  18. Wang L, Zhao W, Wei Z, Liu J (2022) Simkgc: Simple contrastive knowledge graph completion with pre-trained language models. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 4281–4294

  19. Qiu X, Sun T, Xu Y, Shao Y, Dai N, Huang X (2020) Pre-trained models for natural language processing: A survey. Sci China Technol Sci 63(10):1872–1897

    Article  Google Scholar 

  20. Nayyeri M, Wang Z, Akter M, Alam MM, Rony MA, Lehmann J, Staab S, et al. (2022) Integrating knowledge graph embedding and pretrained language models in hypercomplex spaces. arXiv:2208.02743

  21. Li Q, Wang D, Song SFK, Zhang Y, Yu G (2022) Oerl: Enhanced representation learning via open knowledge graphs. IEEE Trans Knowl Data Eng

  22. Bollacker K, Evans C, Paritosh P, Sturge T, Taylor J (2008) Freebase: A collaboratively created graph database for structuring human knowledge. In Proceedings of the 2008 ACM SIGMOD international conference on Management of data, pages 1247–1250

  23. Toutanova K, Chen D (2015) Observed versus latent features for knowledge base and text inference. In Proceedings of the 3rd workshop on continuous vector space models and their compositionality, pages 57–66

  24. Miller GA (1995) Wordnet: a lexical database for english. Commun ACM 38(11):39–41

    Article  Google Scholar 

  25. Shen J, Wang C, Gong L, Song D (2022) Joint language semantic and structure embedding for knowledge graph completion. In Proceedings of the 29th International Conference on Computational Linguistics, pages 1965–1978

  26. Li W, Peng R, Li Z (2022) Improving knowledge graph completion via increasing embedding interactions. Appl Intell 1–19

  27. Zhang S, Sun Z, Zhang W (2020) Improve the translational distance models for knowledge graph embedding. J Intell Inf Syst 55(3):445–467

    Article  Google Scholar 

  28. Li R, Zhao J, Li C, He D, Wang Y, Liu Y, Sun H, Wang S, Deng W, Shen Y, et al. (2022) House: Knowledge graph embedding with householder parameterization. In International Conference on Machine Learning, pages 13209–13224. PMLR

  29. Householder AS (1958) Unitary triangularization of a nonsymmetric matrix. J ACM (JACM) 5(4):339–342

    Article  MathSciNet  MATH  Google Scholar 

  30. Jiang D, Wang R, Yang J, Xue L (2021) Kernel multi-attention neural network for knowledge graph embedding. Knowl-Based Syst 227:107188

    Article  Google Scholar 

  31. Guo L, Sun Z, Hu W (2019) Learning to exploit long-term relational dependencies in knowledge graphs. In International Conference on Machine Learning, pages 2505–2514. PMLR

  32. Wang Q, Huang P, Wang H, Dai S, Jiang W, Liu J, Lyu Y, Zhu Y, Wu H (2019) Coke: Contextualized knowledge graph embedding. arXiv:1911.02168

  33. Youn J, Tagkopoulos I (2022) Kglm: Integrating knowledge graph structure in language models for link prediction. arXiv:2211.02744

  34. Liu Y, Sun Z, Li G, Hu W (2022) I know what you do not know: Knowledge graph embedding via co-distillation learning. In Proceedings of the 31st ACM International Conference on Information & Knowledge Management, pages 1329–1338

  35. Bhargava P, Ng V (2022) Commonsense knowledge reasoning and generation with pre-trained language models: a survey. In Proc AAAI Conf Art Intell 36:12317–12325

    Google Scholar 

  36. Li M, Wang B, Jiang J (2021) Siamese pre-trained transformer encoder for knowledge base completion. Neural Process Lett 53(6):4143–4158

    Article  Google Scholar 

  37. Devlin J, Chang MW, Lee K, Toutanova K (2019) Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of NAACL-HLT, pages 4171–4186

  38. Ji G, He S, Xu L, Liu K, Zhao J (2015) Knowledge graph embedding via dynamic mapping matrix. In Proceedings of the 53rd annual meeting of the association for computational linguistics and the 7th international joint conference on natural language processing (volume 1: Long papers), pages 687–696

  39. Vig J (2019) A multiscale visualization of attention in the transformer model. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics: System Demonstrations, pages 37–42

  40. Van der Maaten L, Hinton G (2008) Visualizing data using t-sne. J Mach Learn Res 9(11)

Download references

Acknowledgements

This work was supported by the National Key Research and Development Program of China (grant number 2021YFC3300204).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Bo Shen.

Ethics declarations

Competing Interests

The authors declare that they have no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Wang, T., Shen, B. & Zhong, Y. SSKGE: a time-saving knowledge graph embedding framework based on structure enhancement and semantic guidance. Appl Intell 53, 25171–25183 (2023). https://doi.org/10.1007/s10489-023-04896-8

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10489-023-04896-8

Keywords

Navigation