/MT-BERT

One Teacher is Enough? Pre-trained Language Model Distillation from Multiple Teachers

No issues in this repository yet.