This repository stores the code of the data augmentation method from Chinese word and character levels, which adds noise to words and characters in redundant, missing, selection and ordering respectively.
pip install -r requirements.txt
Due to copyright restrictions, the two folders have not been uploaded. You can download them from these two links below if you need them.
ChineseHomophones | SimilarCharacter
I've implemented the 5 noise functions described in the paper:
- Delete words with given probability (default is 0.163)
- Replace words by a similar words and homophone with given probability (default is 0.163)
- Swap words up to a certain range (default range is 0.163)
- Repeat words with given probability (default is 0.163)
- Select any of the above noise functions at random
I set the error rate of each time as 16.3%, which can maintain the error rate of the corpus after double noise at 30% (calculated according to mathematical expectation).
Example of simple usage
bash noise.sh
Example of complete usage
python add_noise/add_noise.py --input input_file_path --redundant_probability 0.2 --selection_probability 0.2 --missing_probability 0.2 --ordering_probability 0.2 --comprehensive_error_probability 0.2
I've run Chinese grammatical error correction experiments on Chinese Wikipedia corpus, using all available parallel data.
I added noise to it using this repo, giving the following results on NLPCC 2018 test set. All results are
The table below reports a Transformer model identical to the "base model" in Vaswani et al. (2017).
Model |
|
---|---|
baseline | 33.17 |
baseline+noise | 35.55 |
Transformer base model
@article{xia2022chinese,
title={Chinese grammatical error correction based on knowledge distillation},
author={Xia, Peng and Zhou, Yuechi and Zhang, Ziyan and Tang, Zecheng and Li, Juntao},
journal={arXiv preprint arXiv:2208.00351},
year={2022}
}
@misc{Xia2022ChineseNoisyText,
author = {Peng Xia},
title = {Chinese-Noisy-Text},
year = {2022},
version = {doi},
doi = {10.5281/zenodo.7025129},
publisher = {GitHub},
journal = {GitHub repository},
howpublished = {\url{https://github.com/Richard88888/Chinese-Noisy-Text}}
}
@inproceedings{tang2021基于字词粒度噪声数据增强的中文语法纠错,
title={基于字词粒度噪声数据增强的中文语法纠错 (Chinese Grammatical Error Correction enhanced by Data Augmentation from Word and Character Levels)},
author={Tang, Zecheng and Ji, Yixin and Zhao, Yibo and Li, Junhui},
booktitle={Proceedings of the 20th Chinese National Conference on Computational Linguistics},
pages={813--824},
year={2021}
}
Do not hesitate to contact me if you need some help, need a feature or see some bug
Feel free and welcome to contribute