Some weights of the model checkpoint at ../berts/bert/pytorch_model.bin were not used when initializing WCBertCRFForTokenClassification
bultiful opened this issue · 2 comments
bultiful commented
liuwei1206 commented
Hi,
This is normal. The BERT checkpoint is the parameters of the original BERT trained by masked language model task and next sentence prediction task. For downstream tasks, we only use the encoder of BERT and do not need the layers for mask word prediction and next sentence prediction. Consequently, some weights will not be used for downstream tasks.
Best
bultiful commented
thx u