A Pipeline Of Bert pretraining techniques On Google TPU-pods
A pipeline of various pertaining Bert approaches on text dataset using google TPU.
Introduction
This repo introduces various pre-training approaches to enhance generic BERT models. All engineering proceses were designed for TPU devices. Pre-training methods were implemented using NPL library from TF2 model garden.
- Pre-training from scratch
- Contiunal pre-training
- Simultaneous pre-training
Cite
The study will be published as a publication...
Bibtex
References
[1]Wada, S., Takeda, T., Manabe, S., Konishi, S., Kamohara, J., & Matsumura, Y. (2020). Pre-training technique to localize medical bert and enhance biomedical bert. arXiv preprint arXiv:2005.07202.
Acknowledgements
We would like to acknowledge the TPU Research Cloud program (TRC) and the Google’s CURe program in providing access to TPU-pods and GCP credits, respectively.