/chatbot

Chatbot using Transformer Model and DialoGPT

Primary LanguageJupyter NotebookMIT LicenseMIT

Chatbots

References

Recurrent Neural Networks

In order to understand the concepts of Seq2Seq models or Transformer models, the first thing to know is about RNNs,

Sequence to Sequence Model

A Sequence to Sequence Model aims to map a fixed-length input with a fixed-length output where the length of the input and output may differ.

Transformer Model (Seq2Seq with Attention)

The Transformer in NLP is a novel architecture that aims to solve sequence-to-sequence tasks while handling long-range dependencies with ease.

“The Transformer is the first transduction model relying entirely on self-attention to compute representations of its input and output without using sequence-aligned RNNs or convolution.”

Further Resources