From ddf81c8c0daec9dbac7f3da2bb1a74206e98bba6 Mon Sep 17 00:00:00 2001 From: Shun Kiyono Date: Wed, 18 Sep 2019 19:24:09 +0900 Subject: [PATCH] Added new result for GEC (#345) --- english/grammatical_error_correction.md | 2 ++ 1 file changed, 2 insertions(+) diff --git a/english/grammatical_error_correction.md b/english/grammatical_error_correction.md index 0f917b18..bf05d28a 100644 --- a/english/grammatical_error_correction.md +++ b/english/grammatical_error_correction.md @@ -18,6 +18,7 @@ The shared task setting restricts that systems use only publicly available datas | Model | F0.5 | Paper / Source | Code | | ------------- | :-----:| --- | :-----: | +| Transformer + Pre-train with Pseudo Data (Kiyono et al., EMNLP 2019) | 65.0 | [An Empirical Study of Incorporating Pseudo Data into Grammatical Error Correction](https://arxiv.org/abs/1909.00502) | NA | | Copy-Augmented Transformer + Pre-train (Zhao and Wang, NAACL 2019) | 61.15 | [Improving Grammatical Error Correction via Pre-Training a Copy-Augmented Architecture with Unlabeled Data](https://arxiv.org/pdf/1903.00138.pdf) | [Official](https://github.com/zhawe01/fairseq-gec) | | CNN Seq2Seq + Quality Estimation (Chollampatt and Ng, EMNLP 2018) | 56.52 | [Neural Quality Estimation of Grammatical Error Correction](http://aclweb.org/anthology/D18-1274) | [Official](https://github.com/nusnlp/neuqe/) | | SMT + BiGRU (Grundkiewicz and Junczys-Dowmunt, 2018) | 56.25 | [Near Human-Level Performance in Grammatical Error Correction with Hybrid Machine Translation](http://aclweb.org/anthology/N18-2046)| NA | @@ -116,6 +117,7 @@ Since current state-of-the-art systems rely on as much annotated learner data as | Model | F0.5 | Paper / Source | Code | | ------------- | :-----:| --- | :-----: | +| Transformer + Pre-train with Pseudo Data (Kiyono et al., EMNLP 2019) | 70.2 | [An Empirical Study of Incorporating Pseudo Data into Grammatical Error Correction](https://arxiv.org/abs/1909.00502) | NA | | Transformer | 69.47 | [Neural Grammatical Error Correction Systems with UnsupervisedPre-training on Synthetic Data](https://www.aclweb.org/anthology/W19-4427)| [Official: Code to be updated soon](https://github.com/grammatical/pretraining-bea2019) | | Transformer | 69.00 | [A Neural Grammatical Error Correction System Built OnBetter Pre-training and Sequential Transfer Learning](https://www.aclweb.org/anthology/W19-4423)| [Official](https://github.com/kakaobrain/helo_word/) | | Ensemble of models | 66.78 | [The LAIX Systems in the BEA-2019 GEC Shared Task](https://www.aclweb.org/anthology/W19-4416)| NA |