/BERT-QnA-Squad_2.0_Finetuned_Model

BERT which stands for Bidirectional Encoder Representations from Transformations is the SOTA in Transfer Learning in NLP.

Primary LanguagePython

Watchers