ABSTRACT
The uncertainty is an inherent feature of Knowledge Graph (KG), which is often modelled as confidence scores of relation facts. Although Knowledge Graph Embedding (KGE) has been a great success recently, it is still a big challenge to predict confidence of unseen facts in KG in the continuous vector space. There are several reasons for this situation. First, the current KGE is often concerned with the deterministic knowledge, in which unseen facts’ confidence are treated as zero, otherwise as one. Second, in the embedding space, uncertainty features are not well preserved. Third, approximate reasoning in embedding spaces is often unexplainable and not intuitive. Furthermore, the time and space cost of obtaining embedding spaces with uncertainty preserved are always very high. To address these issues, considering Uncertain Knowledge Graph (UKG), we propose a fast and effective embedding method, UKGsE, in which approximate reasoning and calculation can be quickly performed after generating an Uncertain Knowledge Graph Embedding (UKGE) space in a high speed and reasonable accuracy. The idea is that treating relation facts as short sentences and pre-handling are benefit to the learning and training confidence scores of them. The experiment shows that the method is suitable for the downstream task, confidence prediction of relation facts, whether they are seen in UKG or not. It achieves the best tradeoff between efficiency and accuracy of predicting uncertain confidence of knowledge. Further, we found that the model outperforms state-of-the-art uncertain link prediction baselines on CN15k dataset.
- Federico Bianchi, Gaetano Rossiello, Luca Costabello, Matteo Palmonari, and Pasquale Minervini. 2020. Knowledge Graph Embeddings and Explainable AI. (April 2020). arXiv:2004.14843Google Scholar
- Antoine Bordes, Nicolas Usunier, Alberto Garcia-Durán, Jason Weston, and Oksana Yakhnenko. 2013. Translating Embeddings for Modeling Multi-Relational Data(NIPS’13). Curran Associates Inc., Red Hook, NY, USA, 2787–2795.Google Scholar
- Xuelu Chen, Muhao Chen, Weijia Shi, Yizhou Sun, and Carlo Zaniolo. 2019. Embedding Uncertain Knowledge Graphs. arxiv:1811.10667 [cs.AI]Google Scholar
- Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. (Oct. 2018). arXiv:1810.04805 [cs.CL]Google Scholar
- Felix A. Gers, Jürgen A. Schmidhuber, and Fred A. Cummins. 2000. Learning to Forget: Continual Prediction with LSTM. Neural Comput. 12, 10 (Oct. 2000), 2451–2471.Google ScholarDigital Library
- Shizhu He, Kang Liu, Guoliang Ji, and Jun Zhao. 2015. Learning to Represent Knowledge Graphs with Gaussian Embedding(CIKM ’15). Association for Computing Machinery, New York, NY, USA, 623–632.Google Scholar
- Aidan Hogan, Eva Blomqvist, Michael Cochez, Claudia d’Amato, Gerard de Melo, and et al.2020. Knowledge Graphs. abs/2003.02320 (March 2020). arXiv:2003.02320 [cs.AI]Google Scholar
- Jiafeng Hu, Reynold Cheng, Zhipeng Huang, Yixang Fang, and Siqiang Luo. 2017. On Embedding Uncertain Graphs(CIKM ’17). Association for Computing Machinery, New York, NY, USA, 157–166.Google Scholar
- Tie-Yan Liu. 2010. Learning to Rank for Information Retrieval(SIGIR ’10). Association for Computing Machinery, New York, NY, USA, 904.Google Scholar
- Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013. Efficient Estimation of Word Representations in Vector Space. (Jan. 2013). arXiv:1301.3781Google Scholar
- Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013. Distributed Representations of Words and Phrases and their Compositionality. (Oct. 2013). arXiv:1310.4546Google Scholar
- Mojtaba Nayyeri, Chengjin Xu, Jens Lehmann, and Hamed Shariat Yazdi. 2019. LogicENN: A Neural Based Knowledge Graphs Embedding Model with Logical Rules. (Aug. 2019). arXiv:1908.07141 [cs.AI]Google Scholar
- Mojtaba Nayyeri, Chengjin Xu, Jens Lehmann, and Hamed Shariat Yazdi. 2019. LogicENN: A Neural Based Knowledge Graphs Embedding Model with Logical Rules. arxiv:1908.07141 [cs.AI]Google Scholar
- Radim Řehůřek and Petr Sojka. 2010. Software Framework for Topic Modelling with Large Corpora. ELRA, Valletta, Malta, 45–50.Google Scholar
- Tommaso Soru, Stefano Ruberto, Diego Moussallem, André Valdestilhas, Alexander Bigerl, Edgard Marx, and Diego Esteves. 2018. Expeditious Generation of Knowledge Graph Embeddings. arxiv:1803.07828 [cs.CL]Google Scholar
- Robyn Speer, Joshua Chin, and Catherine Havasi. 2018. ConceptNet 5.5: An Open Multilingual Graph of General Knowledge. arxiv:1612.03975 [cs.CL]Google Scholar
- Luke Vilnis, Xiang Li, Shikhar Murty, and Andrew McCallum. 2018. Probabilistic Embedding of Knowledge Graphs with Box Lattice Measures. (May 2018). arXiv:1805.06627Google Scholar
Recommendations
High accuracy and geometry-consistent confidence prediction network for multi-view stereo
Highlights- A high accuracy and geometry-consistent confidence prediction network that globally fuses both spatial coherence and cross-view consistency.
Graphical abstractDisplay Omitted
AbstractConfidence prediction task attempts to infer the correctness of estimated depth hypotheseshich has gained popularity recently in stereo matching and boosts the accuracy of disparity estimation. However, less attention is paid on ...
Approximate inferring with confidence predicting based on uncertain knowledge graph embedding
AbstractUncertainty is a natural character of knowledge, while it is still tough to be encoded into the knowledge graph embedding space that can be employed for machine learning tasks. However, the approximate inference could be performed in ...
Knowledge graph embedding via multiplicative interaction
ICIAI '18: Proceedings of the 2nd International Conference on Innovation in Artificial IntelligenceKnowledge graphs are playing a crucial role in many machine learning applications. Since most of the knowledge graphs are far from complete, many knowledge graph completion models have been proposed. TransE and its extended models all model knowledge ...
Comments