Neural Bi-Lexicalized PCFG Induction
Songlin Yang, Yanpeng Zhao, Kewei Tu
Abstract
Neural lexicalized PCFGs (L-PCFGs) have been shown effective in grammar induction. However, to reduce computational complexity, they make a strong independence assumption on the generation of the child word and thus bilexical dependencies are ignored. In this paper, we propose an approach to parameterize L-PCFGs without making implausible independence assumptions. Our approach directly models bilexical dependencies and meanwhile reduces both learning and representation complexities of L-PCFGs. Experimental results on the English WSJ dataset confirm the effectiveness of our approach in improving both running speed and unsupervised parsing performance.
Results
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Constituency Parsing | PTB Diagnostic ECG Database | Mean F1 (WSJ) | 60.4 | NBL-PCFG |
Related Papers
On Eliciting Syntax from Language Models via Hashing2024-10-05Improving Unsupervised Constituency Parsing via Maximizing Semantic Information2024-10-03Structural Optimization Ambiguity and Simplicity Bias in Unsupervised Neural Grammar Induction2024-07-23Generative Pretrained Structured Transformers: Unsupervised Syntactic Language Models at Scale2024-03-13Simple Hardware-Efficient PCFGs with Independent Left and Right Productions2023-10-23Ensemble Distillation for Unsupervised Constituency Parsing2023-10-03Augmenting Transformers with Recursively Composed Multi-grained Representations2023-09-28Dynamic Programming in Rank Space: Scaling Structured Inference with Low-Rank HMMs and PCFGs2022-05-01