AristotelisPap/Question-Answering-with-BERT-and-Knowledge-Distillation
Fine-tuned BERT on SQuAd 2.0 Dataset. Applied Knowledge Distillation (KD) and fine-tuned DistilBERT (student) using BERT as the teacher model. Reduced the size of the original BERT by 40%.
Jupyter NotebookMIT
Stargazers
- ahmedamro
- akanametov
- barana91
- CizChenzhoUbbtree
- Damephena
- Hardik-Dharmik-94
- hockeybro12San Francisco, CA
- Jayesh-Kumpawat
- k-praveen-trellisTrellisSoft
- KimGaHyeon12
- luziyi1998
- lymperop
- nazim1021Roche
- nd1511European Space Agency (ESA)
- newlightlw
- penguinwang96825Sheffield, United Kingdom
- pkyriakis
- ScottishFold007Shanghai
- ShuMengZ
- subhasisjAlef Education
- subhobrata
- SunsetCollector
- tikitaka-ball
- tongtz
- Wxhxh33