Adaptive knowledge subgraph ensemble for robust and trustworthy knowledge graph completion

Guojia Wan, Bo Du, Shirui Pan, Jia Wu

Research output: Contribution to journalArticle

Abstract

Knowledge graph (KG) embedding approaches are widely used to infer underlying missing facts based on intrinsic structure information. However, the presence of noisy facts in automatically extracted or crowdsourcing KGs significantly reduces the reliability of various embedding learners. In this paper, we thoroughly study the underlying reasons for the performance drop in dealing with noisy knowledge graphs, and we propose an ensemble framework, Adaptive Knowledge Subgraph Ensemble (AKSE), to enhance the robustness and trust of knowledge graph completion. By employing an effective knowledge subgraph extraction approach to re-sample the sub-components from the original knowledge graph, AKSE generates different representations for learning diversified base learners (e.g., TransE and DistMult), which substantially alleviates the noise effect of KG embedding. All embedding learners are integrated into a unified framework to reduce generalization errors via our simple or adaptive weighting schemes, where the weight is allocated based on each individual learner's prediction capacity. Experimental results show that the robustness of our ensemble framework outperforms exiting knowledge graph embedding approaches on manually injected noise as well as inherent noisy extracted KGs.

Original languageEnglish
Pages (from-to)471-490
Number of pages20
JournalWorld Wide Web
Volume23
Issue number1
DOIs
Publication statusPublished - Jan 2020

Keywords

  • Trustworthy knowledge graph
  • Knowledge graph completion
  • Link prediction
  • Knowledge graph embedding
  • Never-ending language learning

Cite this