/Chaii-2021

Primary LanguageJupyter Notebook

Chaii-2021

This repositiory contains work done by us while trying for the Chaii-2021 challenge. For the challenge we have to design a model to predict answers for questions from articles written in Hindi and Tamil. To make this model we first started to learn and explore concepts about NLP and NLU. We learnt about transformers expecially BERT.
We went through various articles and other resources to get an idea on how to proceed with this problem. Some of the really helpful articles were:

We thought before moving on to Hindi and Tamil we should first try and make a model to predict answers to questions in English. So we used distilBERT( for faster training) to make a model to predict answers to questions from the Stanford question and answering dataset. This train-v2.0.json file is the training dataset from SQUAD that we used. The bert_qna.ipynb has the model that we implemented.

We also tried to get word embeddings for hindi and tamil languages using BERT's word piece tokenizer itself. We directly Chaii 2021's dataset to train this on all the contexts given in the file. The file Hindi_and_Tamil_Tokenization.ipynb contains this implementation.

However we are also exploring google's Electra to use for pre training and generating word embeddings for Hindi and Tamil hoping to get better results.