TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Rethinking Self-Attention: Towards Interpretability in Neu...

Rethinking Self-Attention: Towards Interpretability in Neural Parsing

Khalil Mrini, Franck Dernoncourt, Quan Tran, Trung Bui, Walter Chang, Ndapa Nakashole

2019-11-10Findings of the Association for Computational Linguistics 2020Constituency ParsingDependency Parsing
PaperPDFCodeCode(official)

Abstract

Attention mechanisms have improved the performance of NLP tasks while allowing models to remain explainable. Self-attention is currently widely used, however interpretability is difficult due to the numerous attention distributions. Recent work has shown that model representations can benefit from label-specific information, while facilitating interpretation of predictions. We introduce the Label Attention Layer: a new form of self-attention where attention heads represent labels. We test our novel layer by running constituency and dependency parsing experiments and show our new model obtains new state-of-the-art results for both tasks on both the Penn Treebank (PTB) and Chinese Treebank. Additionally, our model requires fewer self-attention layers compared to existing work. Finally, we find that the Label Attention heads learn relations between syntactic categories and show pathways to analyze errors.

Results

TaskDatasetMetricValueModel
Dependency ParsingPenn TreebankLAS96.26Label Attention Layer + HPSG + XLNet
Dependency ParsingPenn TreebankPOS97.3Label Attention Layer + HPSG + XLNet
Dependency ParsingPenn TreebankUAS97.42Label Attention Layer + HPSG + XLNet
Constituency ParsingPenn TreebankF1 score96.38Label Attention Layer + HPSG + XLNet

Related Papers

Automatic Extraction of Clausal Embedding Based on Large-Scale English Text Data2025-06-16Step-by-step Instructions and a Simple Tabular Output Format Improve the Dependency Parsing Accuracy of LLMs2025-06-11UD-KSL Treebank v1.3: A semi-automated framework for aligning XPOS-extracted units with UPOS tags2025-06-10LKD-KGC: Domain-Specific KG Construction via LLM-driven Knowledge Dependency Parsing2025-05-30Dependency Parsing is More Parameter-Efficient with Normalization2025-05-26FiLLM -- A Filipino-optimized Large Language Model based on Southeast Asia Large Language Model (SEALLM)2025-05-25CrosGrpsABS: Cross-Attention over Syntactic and Semantic Graphs for Aspect-Based Sentiment Analysis in a Low-Resource Language2025-05-25Semantic-based Unsupervised Framing Analysis (SUFA): A Novel Approach for Computational Framing Analysis2025-05-21