Tao Ge, Furu Wei, Ming Zhou
Neural sequence-to-sequence (seq2seq) approaches have proven to be successful in grammatical error correction (GEC). Based on the seq2seq framework, we propose a novel fluency boost learning and inference mechanism. Fluency boosting learning generates diverse error-corrected sentence pairs during training, enabling the error correction model to learn how to improve a sentence's fluency from more instances, while fluency boosting inference allows the model to correct a sentence incrementally with multiple inference steps. Combining fluency boost learning and inference with convolutional seq2seq models, our approach achieves the state-of-the-art performance: 75.72 (F_{0.5}) on CoNLL-2014 10 annotation dataset and 62.42 (GLEU) on JFLEG test set respectively, becoming the first GEC system that reaches human-level performance (72.58 for CoNLL and 62.37 for JFLEG) on both of the benchmarks.
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Grammatical Error Correction | Unrestricted | F0.5 | 61.34 | CNN Seq2Seq + Fluency Boost |
| Grammatical Error Correction | Unrestricted | GLEU | 62.37 | CNN Seq2Seq + Fluency Boost and inference |