TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Structure-aware Fine-tuning of Sequence-to-sequence Transf...

Structure-aware Fine-tuning of Sequence-to-sequence Transformers for Transition-based AMR Parsing

Jiawei Zhou, Tahira Naseem, Ramón Fernandez Astudillo, Young-suk Lee, Radu Florian, Salim Roukos

2021-10-29EMNLP 2021 11AMR Parsing
PaperPDFCode(official)

Abstract

Predicting linearized Abstract Meaning Representation (AMR) graphs using pre-trained sequence-to-sequence Transformer models has recently led to large improvements on AMR parsing benchmarks. These parsers are simple and avoid explicit modeling of structure but lack desirable properties such as graph well-formedness guarantees or built-in graph-sentence alignments. In this work we explore the integration of general pre-trained sequence-to-sequence language models and a structure-aware transition-based approach. We depart from a pointer-based transition system and propose a simplified transition set, designed to better exploit pre-trained language models for structured fine-tuning. We also explore modeling the parser state within the pre-trained encoder-decoder architecture and different vocabulary strategies for the same purpose. We provide a detailed comparison with recent progress in AMR parsing and show that the proposed parser retains the desirable properties of previous transition-based approaches, while being simpler and reaching the new parsing state of the art for AMR 2.0, without the need for graph re-categorization.

Results

TaskDatasetMetricValueModel
Semantic ParsingLDC2017T10Smatch84.7StructBART-J (IBM)
AMR ParsingLDC2017T10Smatch84.7StructBART-J (IBM)

Related Papers

Reassessing Graph Linearization for Sequence-to-sequence AMR Parsing: On the Advantages and Limitations of Triple-Based Encoding2025-05-13Should Cross-Lingual AMR Parsing go Meta? An Empirical Assessment of Meta-Learning and Joint Learning AMR Parsing2024-10-04Adapting Abstract Meaning Representation Parsing to the Clinical Narrative -- the SPRING THYME parser2024-05-15AMR Parsing is Far from Solved: GrAPES, the Granular AMR Parsing Evaluation Suite2023-12-06AMR Parsing with Causal Hierarchical Attention and Pointers2023-10-18Guiding AMR Parsing with Reverse Graph Linearization2023-10-13Incorporating Graph Information in Transformer-based AMR Parsing2023-06-23AMRs Assemble! Learning to Ensemble with Autoregressive Models for AMR Parsing2023-06-19