/Visual-Question-Answering-with-Keras

图片特征和文本特征分别提取

Primary LanguageJupyter Notebook

Visual Question Answering on the FloodNet Dataset

As a step towards learning more about multimodal systems, I took part in the FloodNet Challenge @ EARTHVISION 2021 - Track 2. The track involved developing a Visual Question Answering (VQA) algorithm that could effectively answer questions based on the FloodNet Dataset.


Image Credits: arXiv:2012.02951v1

In my work, I built a simple Joint Embedding VQA-based model, taking inspiration from Akshay Chavan's articles and Github repository.

For image feature extraction, I experimented with two models—VGG16 and InceptionResNetv2. For textual features, I explored the usage of RNNs and LSTMs while attempting to use self-attention.

The model gave an overall accuracy 0.4254 which could be improved by doing the following:

  • Training the model for a greater number of epochs or till it converges.
  • Using attention based models for image features, text features or both.
  • Implementing other concatenation techniques such as Multimodal Compact Bilinear Pooling (MCBP)

The challenge was a first for me as I attempted to use the knowledge I gained from months of reading about VQA systems to finally working on implementing them. As a way forward, I plan to improve my understanding of language models and attention mechanisms before working on more multimodal-based projects.