BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Opened this issue · 0 comments
kweonwooj commented
Abstract
- introduce new language representation model called
BERT
(Bidirectional Encoder Representation from Transformers) - obtains SoTA on GLEU benchmark, MultiNLI and SQuAD v1.1 (almost all natural language inference tasks)
- successful transfer learning in natural language area
Details
Pre-Training (Transfer Learning)
- BERT is the best pre-training model for natural language understanding tasks
- Transfer learning is proven to be useful in many deep learning tasks, especially on image tasks, where ImageNet pre-trained models allow faster training and better performance.
- There are two approaches to utilize transfer learning
feature-based
approach uses extracted feature representation of pre-trained model to downstream taskfine-tuning
approach further trains pre-trained model together with much lower learning rate and much shorter epochs
ELMo vs OpenAI GPT vs BERT
- Classic pre-training model architectures in natural language tasks
ELMo
: use concatenation of bidirectional LSTM representationOpenAI GPT
: use left-to-right Transformer model (Transformer decoder)BERT
: use bidirectional self-attention Transformer model (Transformer encoder)
BERT
Input Representation
- first token is always
[CLS]
(short for classification token) - each token embedding is added with segment embedding (which is sentence embedding pre-trained)
- since BERT will deal with multiple sentence input, segment embedding is introduced by authors
- positional embedding is also added
Pre-training BERT
- not only the model architecture, but the novel pre-training scheme is a big contribution
Task 1 : Masked LM
- traditional left-to-right model has been trained with left-to-right LM, where models were trained to predict the next token
- BERT is trained on masked LM where random input tokens are masked, and model is trained to predict the masked token.
- 15% of the token is masked. Within 15%, 80% is masked as usual, 10% is replaced with random word, 10% is left as-is.
- random masking allows the model to embed all the sentence information in each token (especially in
CLS
token)
Task 2 : Next Sentence Prediction
- in order to align the pre-training with downstream task, add another pre-training task
- input two sentences with separator
[SEP]
and predict whether the sentence is actually adjacent sentence or not.- performs 97~98% accurate (which means the next sentence is easy to classify)
Fine-Tuning in BERT
Dataset (tasks)
- GLEU benchmark is free of evaluation bias. can be classified into below tasks
- sentence-pair classification task : MNLI, QQP, QNLI, STS-B, MRPC, RTE, SWAG
- single sentence classification task : SST-2, CoLA
- QnA task : SQuAD v1.1
- single sentence tagging task : CoNLL-2003 NER
Fine-Tuning
- Sentence-pair classification task and single sentence classification task use final output of
[CLS]
as input to additional FFN layer and softmax (a and b in below figure) - QnA task has two additional vectors to learn (
S
for start andE
for end vector). EachS
andE
is dot-producted with final representation of paragraph tokens to classify which token is starting token and which is end token. (c in below figure) - single sentence tagging tasks uses all final output of each tokens as input to additional FFN layer and softmax model independently, because each output token already has context information, there is no need for model to attend to nearby tokens. Can make non-autoregressive inference on NER task. (d in below figure)
Performance
Ablation Study
What contributed the most?
- People may attribute the success to BERT being bidirectional compared to OpenAPI GPT and ELMo
- To prove that novel pre-training method was also significant, ablation study is performed.
Effect on Model Size
Effect of Training Steps
- huge data and long pre-training is necessary
- although pre-training does take longer than LTR, gain is much higher
Feature-based vs Fine-Tuning
- how well does feature-based transfer learning play in BERT?
Personal Thoughts
- Well-engineered pre-training scheme. Aligns with downstream tasks and shows better performance. Adding both masked LM and next sentence representation is a creative idea.
- Using
[CLS]
as sentence representation was also impressive. The connection between BERT and downstream tasks (Figure 3) is well-designed. - This paper definitely has its contribution in designing novel pre-training methods and naturally linking model with downstream tasks.
- wonder how it can improve Machine Translation?
- For example in EnDe, pre-train English BERT and pre-train German OpenAI GPT and fine-tune using parallel corpus?
Link : https://arxiv.org/pdf/1810.04805.pdf
Authors : Devlin et al 2018