TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/How Does Knowledge Graph Embedding Extrapolate to Unseen D...

How Does Knowledge Graph Embedding Extrapolate to Unseen Data: A Semantic Evidence View

Ren Li, Yanan Cao, Qiannan Zhu, Guanqun Bi, Fang Fang, Yi Liu, Qian Li

2021-09-24Knowledge Graph EmbeddingKnowledge Graph CompletionGraph EmbeddingLink Prediction
PaperPDFCode(official)

Abstract

Knowledge Graph Embedding (KGE) aims to learn representations for entities and relations. Most KGE models have gained great success, especially on extrapolation scenarios. Specifically, given an unseen triple (h, r, t), a trained model can still correctly predict t from (h, r, ?), or h from (?, r, t), such extrapolation ability is impressive. However, most existing KGE works focus on the design of delicate triple modeling function, which mainly tells us how to measure the plausibility of observed triples, but offers limited explanation of why the methods can extrapolate to unseen data, and what are the important factors to help KGE extrapolate. Therefore in this work, we attempt to study the KGE extrapolation of two problems: 1. How does KGE extrapolate to unseen data? 2. How to design the KGE model with better extrapolation ability? For the problem 1, we first discuss the impact factors for extrapolation and from relation, entity and triple level respectively, propose three Semantic Evidences (SEs), which can be observed from train set and provide important semantic information for extrapolation. Then we verify the effectiveness of SEs through extensive experiments on several typical KGE methods. For the problem 2, to make better use of the three levels of SE, we propose a novel GNN-based KGE model, called Semantic Evidence aware Graph Neural Network (SE-GNN). In SE-GNN, each level of SE is modeled explicitly by the corresponding neighbor pattern, and merged sufficiently by the multi-layer aggregation, which contributes to obtaining more extrapolative knowledge representation. Finally, through extensive experiments on FB15k-237 and WN18RR datasets, we show that SE-GNN achieves state-of-the-art performance on Knowledge Graph Completion task and performs a better extrapolation ability. Our code is available at https://github.com/renli1024/SE-GNN.

Results

TaskDatasetMetricValueModel
Link PredictionWN18RRHits@10.446SE-GNN (ours)
Link PredictionWN18RRHits@100.572SE-GNN (ours)
Link PredictionWN18RRHits@30.509SE-GNN (ours)
Link PredictionWN18RRMR3211SE-GNN (ours)
Link PredictionWN18RRMRR0.484SE-GNN (ours)
Link PredictionFB15k-237Hits@10.271SE-GNN (ours)
Link PredictionFB15k-237Hits@100.549SE-GNN (ours)
Link PredictionFB15k-237Hits@30.399SE-GNN (ours)
Link PredictionFB15k-237MR157SE-GNN (ours)
Link PredictionFB15k-237MRR0.365SE-GNN (ours)

Related Papers

SMART: Relation-Aware Learning of Geometric Representations for Knowledge Graphs2025-07-17Topic Modeling and Link-Prediction for Material Property Discovery2025-07-08Graph Collaborative Attention Network for Link Prediction in Knowledge Graphs2025-07-05Understanding Generalization in Node and Link Prediction2025-07-01Context-Driven Knowledge Graph Completion with Semantic-Aware Relational Message Passing2025-06-29Directed Link Prediction using GNN with Local and Global Feature Fusion2025-06-25Demystifying Distributed Training of Graph Neural Networks for Link Prediction2025-06-25Call Me Maybe: Enhancing JavaScript Call Graph Construction using Graph Neural Networks2025-06-22