Divided training sentences into tokens using OpenAI-GPT and Bert tokenizers. Implemented multiple training models: Casual language model, Masked language model, and Sequence-to-sequence model using Pytorch.
tqdat712/Natural-Language-Processing
Using multiple language models to complete sentence and summary paragraph.
Jupyter Notebook