Elsevier

Knowledge-Based Systems

Volume 209, 17 December 2020, 106421
Knowledge-Based Systems

GRL: Knowledge graph completion with GAN-based reinforcement learning

https://doi.org/10.1016/j.knosys.2020.106421Get rights and content

Abstract

Knowledge graph completion intends to infer the entities that need to be queried through the entities and relations known in the knowledge graphs. It is used in many applications, such as question and answer systems, and searching engines. As the completion process can be represented as a Markov process, existing works would solve this problem with reinforcement learning. However, there are three issues blocking them from achieving high accuracy, which are reward sparsity, missing specific domain rules, and ignoring the generation of knowledge graphs. In this paper, we design a generative adversarial net (GAN)-based reinforcement learning model, named GRL, for knowledge graph completion. First, GRL employs the graph convolutional network to embed the knowledge graphs into the low-dimensional space. Second, GRL employs both GAN and long short-term memory (LSTM) to record trajectory sequences obtained by the agent from traversing the knowledge graph and generate new trajectory sequences if needed. At the same time, GRL applies domain-specific rules accordingly. Finally, GRL employs the deep deterministic policy gradient method to optimize both rewards and adversarial loss. The experiments show that GRL is able to both generate better policies and outperform traditional methods for several tasks.

Introduction

Knowledge graph is designed to represent the various entities or concepts and their relations existing in the real world. It constitutes a huge semantic network graph in which nodes represent entities or concepts, and edges represent attributes or relations. Knowledge graph has been used to represent a variety of large-scale knowledge bases. However, the real-world knowledge bases are usually incomplete [1], [2], one needs to complete it by inferring the missing ones with the help of existing information. After that, one can apply the completed knowledge base to real tasks, such as question and answer and searching engines. This knowledge base completion process is known as the knowledge graph completion problem [3], [4], [5].

Existing knowledge graph completion methods can be classified into three types: (1) Path ranking algorithm (PRA) based [6]. They predict the potential relations between entities by connecting existing paths of entities; (2) Representation learning based. They first map the entities and relations into the low-dimensional space, and then reason through vector operations, e.g., TransE [7]; and (3) Probability graph based [8] includes the Markov logic networks and their derivatives, e.g., reinforcement learning methods [9]. These methods have been shown to perform well in their respective areas of expertise.

However, most existing methods did not consider the fact that the knowledge graph can generate new sub-graphs (including nodes and edges). The previous reinforcement learning methods assume that the agent can find the target entity in the knowledge graph within limited time steps [10], [11], [12]. However, the target entity may be missing and thus requires completion. For example, the question “who is the current president of USA?” is formalized as the triple (“USA”, “president of”, ?). If there is a complete triplet (“USA”, “president of”, “Trump”) in the knowledge graph, the agent will find the target entity “Trump” easily. However, if the target entity is missing in the knowledge graph, how could the agent find the correct answer? Inspired by the graph generation techniques [13], we can generate the sub-graphs when the target entity is not found in the known knowledge graph, and let the agent select the correct edges and target entity. Knowledge graph completion is essentially a problem of generating missing parts based on the existing. Compared with the previous methods, we also inferred the missing parts with the help of existing ones in the knowledge graph, but we use the graph generation techniques to strengthen the generation ability and optimize the reward so that the agent can find the target entities or missing entities more intelligently. Further, in existing reinforcement learning based models, the sampling method faces large variance and relatively low learning efficiency, and thus faces the challenges of fragile convergence and sparse rewards.

Deep neural network has shown its great flexibility and power by learning to represent the world as a nested hierarchy of concepts, achieving excellent results in lots of applications [14]. It should be pointed out that when using deep neural networks, very little work directly uses a single neural network to encode the graph structure, but combines multiple different forms of neural networks to enhance the capabilities of the entire network architecture. Motivated by these, in this paper, we propose a new framework based on GAN [15] and reinforcement learning for knowledge graph completion, named GRL. We divide the knowledge graph completion problem into two parts: first, the target entity can be found within the limited time step from the original knowledge graph; second, the target entity cannot be found from the original knowledge graph while there are still time steps left, one can generate a new sub-graph as shown in Fig. 1. In other words, when the agent did not find an answer within the time steps, it can generate a new path at the right time, which refers to the time when the agent thinks that generating a new path or node is more conducive to find the target entity based on the rules. This mechanism is more intelligent than the previous work. As reinforcement learning method learns policies through interaction with the environment to guide the agent to walk in the knowledge graph [10], [11], it is a good fit for sequence decision-making problems. As a result, we define the knowledge graph completion problem as a Markov process, and explore the rules that can be introduced in both state transition process and rewards, which can better guide the walking path under the optimization of GAN. We employ long short-term memory (LSTM) [16] as a generator of GAN, which not only records historical trajectories but also generates new sub-graphs, and trains policy networks with GAN. Further, to better generate new sub-graphs, a graph neural network (GNN) [17] is used to embed the knowledge graph into low-dimensional vectors and parameterize the message passing process at each layer. Finally, we employ DDPG [18] to optimize the entire model.

To summarize, our main contributions are:

  • We are the first to introduce the graph generation techniques to knowledge graph reasoning to help agents better find the target entity.

  • We are the first to use GAN to optimize reward in the knowledge graph completion problem, in which the introduced rules are optimized to guide the agent’s walking path and combined with trajectory generation to train the policy network.

  • DDPG is employed in the framework to solve the problems of sparse rewards and low sampling efficiency brought by REINFORCE in previous work. The experimental results show the efficiency and effectiveness of our model.

Section snippets

Related work

Knowledge graph completion methods are mainly divided into three categories: representation learning based, path ranking based, and reinforcement learning based.

Representation learning aims to represent the semantic information of the research object as a dense low-dimensional real-value vector, while knowledge representation learning [19] is representational learning of entities and relations in the knowledge graph [20]. We can use modeling methods to represent entities and vectors in

Problem definition

We formally define the research problem of this paper in the following. The knowledge graph can be defined as G={E,R}, where E represents the set of all entities (e) and R represents the set of all relations (r) that exist in the knowledge graph or newly generated entities or relations. In the applications such as searching and Q&A, most problems are to infer another entity when we know an entity and relation, which can be formed by the triple in the knowledge graph as (eS,rq,?), where eS

Experiments

In this section, we describe the experiments in detail to verify the effectiveness and efficiency of our model. We first describe the data sets and parameter settings we use in the experiment, and then we design a series of experiments, such as link prediction, fact classification, ablation study, and sensitivity analysis. We prove that our model’s overall performance is better than the traditional embedding-based methods, random walk methods, and other baseline methods, especially in

Conclusion and future work

We propose a novel framework for knowledge graph completion. We observe that the dynamics in the knowledge graph are important factors for knowledge graph completion. We then design a set of a neural network architecture based on GAN and reinforcement learning. We use WGAN to introduce rules to better guide agents to find paths, dynamically record, and generate sub-graph sequences, and form policies in conjunction with LSTM. Finally, we utilize DDPG methods to optimize the policy. Experiments

CRediT authorship contribution statement

Qi Wang, Yuede Ji, Yongsheng Hao and Jie Cao conceived and designed the study, and conducted experiments. Qi Wang wrote the paper, and Yuede Ji, Yongsheng Hao and Cao Jie reviewed and edited the manuscript. All authors read and approve the manuscript.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Acknowledgments

The work was partly supported by the Chinese National Funding of Social Sciences (No. 16ZDA054), the National Natural Science Foundation of China (NSFC) under grant (NO. 41475089).

Qi Wang received the B.S. and M.Eng. degree in Software engineering from Jilin University (Changchun city) and Central South University (Changsha city) in 2012 and 2016, China, respectively. He is currently pursuing the PH.D. degree in Software engineering at the school of computer science, Fudan University, Shanghai, China.

His current research interests include knowledge graph, deep learning, and reinforcement learning.

References (52)

  • Y. Zhang, H. Dai, Z. Kozareva, A.J. Smola, L. Song, Variational reasoning for question answering with knowledge graph,...
  • Variational knowledge graph reasoning

  • TrouillonThéo

    Knowledge graph completion via complex tensor factorization

    J. Mach. Learn. Res.

    (2017)
  • GaoJianfeng et al.

    Neural approaches to conversational ai

    Found. Trends® Inf. Retr.

    (2019)
  • LaoNi et al.

    Relational retrieval using a combination of path-constrained random walks

    Mach. Learn.

    (2010)
  • LaoNi et al.

    Cohen relational retrieval using a combination of path-constrained random walks

    Mach. Learn.

    (2010)
  • BordesAntoine

    Translating embeddings for modeling multi-relational data

    Adv. Neural Inf. Process. Syst.

    (2013)
  • HintonGeoffrey

    Deep neural networks for acoustic modeling in speech recognition: The shared views of four research groups

    IEEE Signal Process. Mag.

    (2012)
  • WangH. et al.

    Incorporating graph attention mechanism into knowledge graph reasoning based on deep reinforcement learning

  • XiongWenhan et al.

    Deeppath: A reinforcement learning method for knowledge graph reasoning

    (2017)
  • DasRajarshi

    Go for a walk and arrive at the answer: Reasoning over paths in knowledge bases using reinforcement learning

    (2017)
  • ShenYelong

    M-walk: Learning to walk over graphs using monte carlo tree search

    Adv. Neural Inf. Process. Syst.

    (2018)
  • HjelmR. Devon

    Boundary-seeking generative adversarial networks

    (2017)
  • YoonKim

    Convolutional neural networks for sentence classification

    (2014)
  • Matthew B.A. McDermott, et al. Semi-supervised biomedical translation with cycle wasserstein regression GANs, in:...
  • Inci M. Baytas, et al. Patient subtyping via time-aware LSTM networks, in: Proceedings of the 23rd ACM SIGKDD...
  • LillicrapTimothy P.

    Continuous control with deep reinforcement learning

    (2015)
  • SuttonRichard S. et al.

    Reinforcement learning: An introduction

    (2018)
  • XieRuobing et al.

    Representation learning of knowledge graphs with hierarchical types

  • YaoL. et al.

    KG-BERT: BERT for knowledge graph Completion

    (2019)
  • MaTinghuai et al.

    Graph classification based on graph set reconstruction and graph kernel feature reduction

    Neurocomputing

    (2018)
  • Zhen Wang, et al. Knowledge graph embedding by translating on hyperplanes, in: Twenty-Eighth AAAI cOnference on...
  • Guoliang Ji, et al. Knowledge graph embedding via dynamic mapping matrix, in: Proceedings of the 53rd Annual Meeting of...
  • Matt Gardner, et al. Improving learning and inference in a large knowledge-base using latent syntactic cues, in:...
  • RongH. et al.

    Deep rolling: A novel emotion prediction model for a multi- participant communication context

    Inform. Sci.

    (2019)
  • Matt Gardner, et al. Incorporating vector space similarity in random walk inference over knowledge bases, in:...
  • Cited by (0)

    Qi Wang received the B.S. and M.Eng. degree in Software engineering from Jilin University (Changchun city) and Central South University (Changsha city) in 2012 and 2016, China, respectively. He is currently pursuing the PH.D. degree in Software engineering at the school of computer science, Fudan University, Shanghai, China.

    His current research interests include knowledge graph, deep learning, and reinforcement learning.

    Yuede Ji is a Ph.D. student from the Department of Electrical and Computer Engineering at The George Washington University advised by Prof. H. Howie Huang. His research interests are mainly in graph analytics for high performance computing, cybersecurity, and machine learning. Prior to joining GWU, He received B.E. and M.S. both from Jilin University.

    Yongsheng Hao received his MS Degree of Engineering from Qingdao University in 2008. Now, he is a senior engineer of Network Center, Nanjing University of Information Science & Technology. His current research interests include distributed and parallel computing, mobile computing, Grid computing, web Service, particle swarm optimization algorithm and genetic algorithm. He has published more than 30 papers in international conferences and journals.

    Jie Cao received the Ph.D. degree from Southeast University, Nanjing, China, in 2005. He was an Associate Professor, from 1999 to 2006. From 2006 to 2009, he was a Postdoctoral Fellow of the Academy of Mathematics and Systems Science, Chinese Academy of Science. From 2009 to 2019, he was a Professor with the School of Management and Economics, Nanjing University of Information Science and Technology. Since May 2019, he has been the Vice President of the Xuzhou University of Technology. His research interests include system engineering, and management science and technology.

    View full text