IntelLabs/academic-budget-bert
Repository containing code for "How to Train BERT with an Academic Budget" paper
PythonApache-2.0
Issues
- 0
Grad overflow and null validation loss
#33 opened by NewDriverLee - 4
Which vocabulary file need to use?
#32 opened by NewDriverLee - 1
- 0
The file produced by process_data.py is empty
#28 opened by Richar-Du - 10
- 1
What is the size of the processed data?
#24 opened by leoozy - 1
the eval_acc on RTE dataset is only 55%
#27 opened by leoozy - 11
GLUE results not reproducible
#18 opened by lumliolum - 3
Distributed pretraining dataset question
#22 opened by sangmichaelxie - 1
Finetuning commands for other glue tasks
#25 opened by raghavlite - 10
Unable to train a roberta model?
#8 opened by dseddah - 1
only test_shard_*.hdf5
#21 opened by shizhediao - 1
GLUE dev results
#17 opened by BaohaoLiao - 1
Question: Easiest way to load deepspeed checkpoints as standard PyTorch models?
#16 opened by QuintinPope - 4
- 2
bert_model not used
#15 opened by senisioi - 1
Which versions for pre-training?
#14 opened by marcelbra - 2
- 0
Unable to run_glue
#10 opened by Rotendahl - 6
Question about validation and testing
#5 opened by peerdavid - 2
Any plan for releasing the checkpoints?
#2 opened by gaotianyu1350 - 2
Code release date?
#1 opened by RyanHuangNLP