This repository is a C++ version of the Python HuggingFace tokenizers.
In the HuggingFace Transformers repo, tokenization is done with 104,603 lines of Python code. It takes 5506 lines for GPT2-specific BPE. I went through the code using the Python Debugger (pdb). It turns out that most of them do nothing but virtual methods in a complicated class hierarchy. I took 120 lines of Python code and put them in the file bpe.py
. This program is not a weaker or simpler version; it is the full HuggingFace BPE tokenizer.
With this extraction, I can port it to C++: bpe.h
, bpe.cc
, and bpe_test.cc
.
The following tech notes might help you understand the Python and C++ code:
- Unicode in C++ and Python
- Understanding HuggingFace Tokenizers
- Unicode-enabled Regular Expression in C++
To run the vanilla HuggingFace GPT-2's BPE tokenizer, follow [this link])https://huggingface.co/docs/transformers/model_doc/gpt2#transformers.GPT2Tokenizer).
To run HuggingFace vanilla BPE tokenzier for GPT2, run the following commands:
pip install transformers
python tool/t.py
Please run the following commands to run the extracted 120-line Python tokenizer:
python tokenizer/bpe.py
To build the C++ port, you will need CMake.
cmake -B /tmp/b -S .
Cmake --build /tmp/b
Then you can run the unit test.
/tmp/b/bin/bpe_test
All three of the above programs load the same vocab.json
and merges.txt
files, so they all work exactly like HuggingFace did when it trained and served the GPT2 model.
RE2 can match Unicode letters from all languages using rules like \p{L}
. But it doesn't work with look-ahead syntax like (?!...)
. This would make the C++ version act a little differently than the Python versions when there are more than one space between two words. Please tell me about any C++ regular expression libraries that can handle both Unicode and look-ahead.
This does not seem a big issue because among the 15,231,221 lines in the lyrics text dataset of 54 languages, the C++ tokenizer generates different output from the HuggingFace one for only 4 lines. And the reason are all due to successive two spaces. For details, please see #10.