TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Integrating a Heterogeneous Graph with Entity-aware Self-a...

Integrating a Heterogeneous Graph with Entity-aware Self-attention using Relative Position Labels for Reading Comprehension Model

Shima Foolad, Kourosh Kiani

2023-07-19Reading ComprehensionCommon Sense ReasoningMachine Reading ComprehensionGraph Attention
PaperPDF

Abstract

Despite the significant progress made by transformer models in machine reading comprehension tasks, they still fall short in handling complex reasoning tasks due to the absence of explicit knowledge in the input sequence. To address this limitation, many recent works have proposed injecting external knowledge into the model. However, selecting relevant external knowledge, ensuring its availability, and requiring additional processing steps remain challenging. In this paper, we introduce a novel attention pattern that integrates reasoning knowledge derived from a heterogeneous graph into the transformer architecture without relying on external knowledge. The proposed attention pattern comprises three key elements: global-local attention for word tokens, graph attention for entity tokens that exhibit strong attention towards tokens connected in the graph as opposed to those unconnected, and the consideration of the type of relationship between each entity token and word token. This results in optimized attention between the two if a relationship exists. The pattern is coupled with special relative position labels, allowing it to integrate with LUKE's entity-aware self-attention mechanism. The experimental findings corroborate that our model outperforms both the cutting-edge LUKE-Graph and the baseline LUKE model across two distinct datasets: ReCoRD, emphasizing commonsense reasoning, and WikiHop, focusing on multi-hop reasoning challenges.

Results

TaskDatasetMetricValueModel
Common Sense ReasoningReCoRDEM91.7GESA 500M
Common Sense ReasoningReCoRDF192.2GESA 500M

Related Papers

Comparing Apples to Oranges: A Dataset & Analysis of LLM Humour Understanding from Traditional Puns to Topical Jokes2025-07-17Catching Bid-rigging Cartels with Graph Attention Neural Networks2025-07-16Wavelet-Enhanced Neural ODE and Graph Attention for Interpretable Energy Forecasting2025-07-14LoSiA: Efficient High-Rank Fine-Tuning via Subnet Localization and Optimization2025-07-06DeRIS: Decoupling Perception and Cognition for Enhanced Referring Image Segmentation through Loopback Synergy2025-07-02Following the Clues: Experiments on Person Re-ID using Cross-Modal Intelligence2025-07-02Temporal-Aware Graph Attention Network for Cryptocurrency Transaction Fraud Detection2025-06-26Accessible Gesture-Driven Augmented Reality Interaction System2025-06-18