Training with Exploration Improves a Greedy Stack-LSTM Parser
Miguel Ballesteros, Yoav Goldberg, Chris Dyer, Noah A. Smith
2016-03-11Dependency Parsing
Abstract
We adapt the greedy Stack-LSTM dependency parser of Dyer et al. (2015) to support a training-with-exploration procedure using dynamic oracles(Goldberg and Nivre, 2013) instead of cross-entropy minimization. This form of training, which accounts for model predictions at training time rather than assuming an error-free action history, improves parsing accuracies for both English and Chinese, obtaining very strong results for both languages. We discuss some modifications needed in order to get training with exploration to work well for a probabilistic neural-network.
Results
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Dependency Parsing | Penn Treebank | LAS | 91.42 | Arc-hybrid |
| Dependency Parsing | Penn Treebank | POS | 97.3 | Arc-hybrid |
| Dependency Parsing | Penn Treebank | UAS | 93.56 | Arc-hybrid |
Related Papers
Step-by-step Instructions and a Simple Tabular Output Format Improve the Dependency Parsing Accuracy of LLMs2025-06-11UD-KSL Treebank v1.3: A semi-automated framework for aligning XPOS-extracted units with UPOS tags2025-06-10LKD-KGC: Domain-Specific KG Construction via LLM-driven Knowledge Dependency Parsing2025-05-30Dependency Parsing is More Parameter-Efficient with Normalization2025-05-26FiLLM -- A Filipino-optimized Large Language Model based on Southeast Asia Large Language Model (SEALLM)2025-05-25CrosGrpsABS: Cross-Attention over Syntactic and Semantic Graphs for Aspect-Based Sentiment Analysis in a Low-Resource Language2025-05-25Semantic-based Unsupervised Framing Analysis (SUFA): A Novel Approach for Computational Framing Analysis2025-05-21Hierarchical Bracketing Encodings for Dependency Parsing as Tagging2025-05-16