Issues
- 1
The exact English pretraining data and Chinese pretraining data that are exact same to the BERT paper's pretraining data.
#11 opened by guotong1988 - 2
Pre-training-Using-Knowledge-Distillation is better than Pre-training-Only for downstream tasks?
#10 opened by guotong1988 - 3
Mask-Filling with pretrained BORT
#9 opened by patrickvonplaten - 1
how to train model on another language?
#8 opened by Archelunch - 2
Huggingface support
#4 opened by sbsky - 1
- 5
bort pretrain
#6 opened by nicexw - 1
I couldn't understand the configuration of the model. please can someone clarify?
#3 opened by preethamgali - 1
Can't download model.
#1 opened by hardfish82 - 1
Can't download model again!
#5 opened by killua-zyk - 1