/attention-translation-keras

Attention based sequence to sequence neural machine translation model built in keras.

Primary LanguagePython

Attention based Language Translation in Keras

This is a Attention based sequence to sequence neural machine translation model built in keras.

The same code can be used for any text based sequence to sequence task such as a chatbot!.

The code has only been tested with the tensorflow backend.

Requirements

  • Python 2
  • Tensorflow
  • numpy
  • keras ( Latest )

Instructions

  1. Dowload the dataset

We need to create two .txt files for each the two languages. Both the text files should contain same number of lines and make sure the lines of the both text files ae syncronzed. Make sure that the sentences are tokenised in the text file.

mkdir data
cd data
wget "http://www.cfilt.iitb.ac.in/iitb_parallel/iitb_corpus_download/parallel.tgz"
tar -xf "parallel.tgz"
cd ..
  1. Preprocess the dataset

Now we need preprocess the dataset into an HDF5 file.

python prep_data.py --text_A="data/parallel/IITB.en-hi.en" --text_B="data/parallel/IITB.en-hi.hi" --out_file="./data/nmt_hi_en_prepped.h5"
  1. Start the training
mkdir weights

python train.py --dataset="./data/nmt_hi_en_prepped.h5" --weights_path="./weights/KerasAttentionNMT_1.h5"
  1. Get the predictions from the model
python predict.py --dataset="./data/nmt_hi_en_prepped.h5" --weights_path="./weights/KerasAttentionNMT_1.h5"

Results

===============
Enter a sentence : 
this is red
यह लाल ह <end> 
===============