SKTBrain/KoBERT

What subword tokenizer do you use?

conan1024hao opened this issue · 2 comments

Hi, I am just curious about that what subword tokenizer did you use when training the model, BPE or WordPiece?

We trained this model using Sentencepiece Unigram tokenizer.

@bage79 감사합니다!