TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/AMR Parsing with Action-Pointer Transformer

AMR Parsing with Action-Pointer Transformer

Jiawei Zhou, Tahira Naseem, Ramón Fernandez Astudillo, Radu Florian

2021-04-29NAACL 2021 4AMR Parsing
PaperPDFCode

Abstract

Abstract Meaning Representation parsing is a sentence-to-graph prediction task where target nodes are not explicitly aligned to sentence tokens. However, since graph nodes are semantically based on one or more sentence tokens, implicit alignments can be derived. Transition-based parsers operate over the sentence from left to right, capturing this inductive bias via alignments at the cost of limited expressiveness. In this work, we propose a transition-based system that combines hard-attention over sentences with a target-side action pointer mechanism to decouple source tokens from node representations and address alignments. We model the transitions as well as the pointer mechanism through straightforward modifications within a single Transformer architecture. Parser state and graph structure information are efficiently encoded using attention heads. We show that our action-pointer approach leads to increased expressiveness and attains large gains (+1.6 points) against the best transition-based AMR parser in very similar conditions. While using no graph re-categorization, our single model yields the second best Smatch score on AMR 2.0 (81.8), which is further improved to 83.4 with silver data and ensemble decoding.

Results

TaskDatasetMetricValueModel
Semantic ParsingLDC2014T12F1 Full78.5APT (IBM)
Semantic ParsingLDC2017T10Smatch82.6APT base (IBM)
Semantic ParsingLDC2020T02Smatch80.4APT+Silver (IBM)
AMR ParsingLDC2014T12F1 Full78.5APT (IBM)
AMR ParsingLDC2017T10Smatch82.6APT base (IBM)
AMR ParsingLDC2020T02Smatch80.4APT+Silver (IBM)

Related Papers

Reassessing Graph Linearization for Sequence-to-sequence AMR Parsing: On the Advantages and Limitations of Triple-Based Encoding2025-05-13Should Cross-Lingual AMR Parsing go Meta? An Empirical Assessment of Meta-Learning and Joint Learning AMR Parsing2024-10-04Adapting Abstract Meaning Representation Parsing to the Clinical Narrative -- the SPRING THYME parser2024-05-15AMR Parsing is Far from Solved: GrAPES, the Granular AMR Parsing Evaluation Suite2023-12-06AMR Parsing with Causal Hierarchical Attention and Pointers2023-10-18Guiding AMR Parsing with Reverse Graph Linearization2023-10-13Incorporating Graph Information in Transformer-based AMR Parsing2023-06-23AMRs Assemble! Learning to Ensemble with Autoregressive Models for AMR Parsing2023-06-19