TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Transition-based Parsing with Stack-Transformers

Transition-based Parsing with Stack-Transformers

Ramon Fernandez Astudillo, Miguel Ballesteros, Tahira Naseem, Austin Blodgett, Radu Florian

2020-10-20Findings of the Association for Computational Linguistics 2020AMR Parsing
PaperPDFCode(official)

Abstract

Modeling the parser state is key to good performance in transition-based parsing. Recurrent Neural Networks considerably improved the performance of transition-based systems by modelling the global state, e.g. stack-LSTM parsers, or local state modeling of contextualized features, e.g. Bi-LSTM parsers. Given the success of Transformer architectures in recent parsing systems, this work explores modifications of the sequence-to-sequence Transformer architecture to model either global or local parser states in transition-based parsing. We show that modifications of the cross attention mechanism of the Transformer considerably strengthen performance both on dependency and Abstract Meaning Representation (AMR) parsing tasks, particularly for smaller models or limited training data.

Results

TaskDatasetMetricValueModel
Semantic ParsingLDC2017T10Smatch79stack-Transformer (IBM)
AMR ParsingLDC2017T10Smatch79stack-Transformer (IBM)

Related Papers

Reassessing Graph Linearization for Sequence-to-sequence AMR Parsing: On the Advantages and Limitations of Triple-Based Encoding2025-05-13Should Cross-Lingual AMR Parsing go Meta? An Empirical Assessment of Meta-Learning and Joint Learning AMR Parsing2024-10-04Adapting Abstract Meaning Representation Parsing to the Clinical Narrative -- the SPRING THYME parser2024-05-15AMR Parsing is Far from Solved: GrAPES, the Granular AMR Parsing Evaluation Suite2023-12-06AMR Parsing with Causal Hierarchical Attention and Pointers2023-10-18Guiding AMR Parsing with Reverse Graph Linearization2023-10-13Incorporating Graph Information in Transformer-based AMR Parsing2023-06-23AMRs Assemble! Learning to Ensemble with Autoregressive Models for AMR Parsing2023-06-19