TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Structure-Augmented Text Representation Learning for Effic...

Structure-Augmented Text Representation Learning for Efficient Knowledge Graph Completion

Bo Wang, Tao Shen, Guodong Long, Tianyi Zhou, Yi Chang

2020-04-30Knowledge GraphsRepresentation LearningKnowledge Graph CompletionGraph EmbeddingLink Prediction
PaperPDFCode(official)

Abstract

Human-curated knowledge graphs provide critical supportive information to various natural language processing tasks, but these graphs are usually incomplete, urging auto-completion of them. Prevalent graph embedding approaches, e.g., TransE, learn structured knowledge via representing graph elements into dense embeddings and capturing their triple-level relationship with spatial distance. However, they are hardly generalizable to the elements never visited in training and are intrinsically vulnerable to graph incompleteness. In contrast, textual encoding approaches, e.g., KG-BERT, resort to graph triple's text and triple-level contextualized representations. They are generalizable enough and robust to the incompleteness, especially when coupled with pre-trained encoders. But two major drawbacks limit the performance: (1) high overheads due to the costly scoring of all possible triples in inference, and (2) a lack of structured knowledge in the textual encoder. In this paper, we follow the textual encoding paradigm and aim to alleviate its drawbacks by augmenting it with graph embedding techniques -- a complementary hybrid of both paradigms. Specifically, we partition each triple into two asymmetric parts as in translation-based graph embedding approach, and encode both parts into contextualized representations by a Siamese-style textual encoder. Built upon the representations, our model employs both deterministic classifier and spatial measurement for representation and structure learning respectively. Moreover, we develop a self-adaptive ensemble scheme to further improve the performance by incorporating triple scores from an existing graph embedding model. In experiments, we achieve state-of-the-art performance on three benchmarks and a zero-shot dataset for link prediction, with highlights of inference costs reduced by 1-2 orders of magnitude compared to a textual encoding method.

Results

TaskDatasetMetricValueModel
Link PredictionUMLSHits@100.991StAR
Link PredictionUMLSMR1.49StAR
Link PredictionWN18RRHits@10.243StAR(Self-Adp)
Link PredictionWN18RRHits@100.709StAR(Self-Adp)
Link PredictionWN18RRHits@30.491StAR(Self-Adp)
Link PredictionWN18RRMR51StAR(Self-Adp)
Link PredictionWN18RRMRR0.401StAR(Self-Adp)
Link PredictionFB15k-237Hits@10.266StAR
Link PredictionFB15k-237Hits@100.562StAR
Link PredictionFB15k-237Hits@30.404StAR
Link PredictionFB15k-237MR117StAR
Link PredictionFB15k-237MRR0.365StAR

Related Papers

Touch in the Wild: Learning Fine-Grained Manipulation with a Portable Visuo-Tactile Gripper2025-07-20SMART: Relation-Aware Learning of Geometric Representations for Knowledge Graphs2025-07-17Spectral Bellman Method: Unifying Representation and Exploration in RL2025-07-17Boosting Team Modeling through Tempo-Relational Representation Learning2025-07-17Similarity-Guided Diffusion for Contrastive Sequential Recommendation2025-07-16Are encoders able to learn landmarkers for warm-starting of Hyperparameter Optimization?2025-07-16Language-Guided Contrastive Audio-Visual Masked Autoencoder with Automatically Generated Audio-Visual-Text Triplets from Videos2025-07-16A Mixed-Primitive-based Gaussian Splatting Method for Surface Reconstruction2025-07-15