Encoder-Decoder Models Can Benefit from Pre-trained Masked Language Models in Grammatical Error Correction
Code for the paper: "Can Encoder-Decoder Models Benefit from Pre-trained Language Representation in Grammatical Error Correction?" (In ACL 2020). If you use any part of this work, make sure you include the following citation:
@inproceedings{Kaneko:ACL:2020,
title={Encoder-Decoder Models Can Benefit from Pre-trained Masked Language Models in Grammatical Error Correction},
author={Masahiro Kaneko, Masato Mita, Shun Kiyono, Jun Suzuki and Kentaro Inui},
booktitle={Proc. of the 58th Annual Meeting of the Association for Computational Linguistics (ACL)},
year={2020}
}
- python >= 3.5
- torch == 1.2.0
- bert-nmt
- subword
- gec-pseudodata
- First download the necessary tools using the following command:
cd scripts
./setup.sh
- This code uses wi+locness dataset.
- Note that since the gold of wi+locnness test data is not available, validatuon data was specified as test data.
- Place your data in the
data
directory if necessary. - You can train the BERT-GEC model with the following command:
./train.sh
- You can also correct your ungrammatical data with the following command:
./generate.sh /path/your/data
See the LICENSE file