antdurrant's Stars
infiniflow/ragflow
RAGFlow is an open-source RAG (Retrieval-Augmented Generation) engine based on deep document understanding.
microsoft/NeuralSpeech
cofe-ai/fast-gector
gbateson/moodle-mod_reader
Reader module for Moodle >= 2.0
gbateson/moodle-qtype_essayautograde
Essay (auto-grade) question type for Moodle >= 3.0
sunilchomal/GECwBERT
Use Language Model (LM) for Grammar Error Correction (GEC), without the use of annotated data.
kanekomasahiro/grammatical-error-detection
neuspell/neuspell
NeuSpell: A Neural Spelling Correction Toolkit
borgr/GEC_UD_divergences
taku-ito/INLG2019_SentRev
AutoTemp/Shallow-Aggressive-Decoding
Codes for the paper "Instantaneous Grammatical Error Correction with Shallow Aggressive Decoding" (ACL-IJCNLP 2021)
mhagiwara/github-typo-corpus
GitHub Typo Corpus: A Large-Scale Multilingual Dataset of Misspellings and Grammatical Errors
kanyun-inc/fairseq-gec
Source code for paper: Improving Grammatical Error Correction via Pre-Training a Copy-Augmented Architecture with Unlabeled Data
michiyasunaga/LM-Critic
[EMNLP 2021] LM-Critic: Language Models for Unsupervised Grammatical Error Correction
Katsumata420/generic-pretrained-GEC
Stronger Baselines for Grammatical Error Correction Using a Pretrained Encoder-Decoder Model.
SimonHFL/CWEB
ufal/wnut2021_character_transformations_gec
The code from the paper Character Transformations for Non-Autoregressive GEC Tagging
ufal/low-resource-gec-wnut2019
Source code for paper Grammatical Error Correction in Low-Resource Scenarios (W-NUT 2019)
glnmario/semchange-profiling
Repository for the CoNLL-2021 paper "Grammatical Profiling for Semantic Change Detection" by Mario Giulianelli, Andrey Kutuzov, and Lidia Pivovarova.
google-research-datasets/C4_200M-synthetic-dataset-for-grammatical-error-correction
This dataset contains synthetic training data for grammatical error correction. The corpus is generated by corrupting clean sentences from C4 using a tagged corruption model. The approach and the dataset are described in more detail by Stahlberg and Kumar (2021) (https://www.aclweb.org/anthology/2021.bea-1.4/)
EricFillion/happy-transformer
Happy Transformer makes it easy to fine-tune and perform inference with NLP Transformer models.
google-research/multilingual-t5
google-research/text-to-text-transfer-transformer
Code for the paper "Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer"
ropenscilabs/tif
Text Interchange Formats
cnap/gec-ranking
Data and code used in the 2015 ACL paper, "Ground Truth for Grammatical Error Correction Metrics"
awasthiabhijeet/PIE
Fast + Non-Autoregressive Grammatical Error Correction using BERT. Code and Pre-trained models for paper "Parallel Iterative Edit Models for Local Sequence Transduction": www.aclweb.org/anthology/D19-1435.pdf (EMNLP-IJCNLP 2019)
chrisjbryant/errant
ERRor ANnotation Toolkit: Automatically extract and classify grammatical errors in parallel original and corrected sentences.
keisks/wmt-trueskill
Data and code used in the 2014 WMT, "Efficient Elicitation of Annotations for Human Evaluation of Machine Translation"
EleutherAI/gpt-neo
An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.
kakaobrain/pororo
PORORO: Platform Of neuRal mOdels for natuRal language prOcessing