/Multimodal-Offensive-Meme-Classification-using-Transformers-and-BiLSTM

This is my work on offensive meme classification. This work is published in an International journal. Paper and code can be found here.

Multimodal-Offensive-Meme-Classification-using-Transformers-and-BiLSTM

This is my work on offensive meme classification. This work is published in an International journal. Paper and code can be found here.

Cite

Roshan Nayak, B S Ullas Kannantha, Kruthi S, C. Gururaj. (2022). Multimodal Offensive Meme Classification using Transformers and BiLSTM. International Journal of Engineering and Advanced Technology (Vol. 11, Issue 3, pp. 96–102) Blue Eyes Intelligence Engineering and Sciences Engineering and Sciences Publication - BEIESP. https://doi.org/10.35940/ijeat.c3392.0211322

Abstract

Nowadays memes have become a way in which people express their ideas on social media. These memes can convey various views including offensive ones. Memes can be intended for a personal attack, homophobic abuse, racial abuse, attack on minority etc. The memes are implicit and multi-modal in nature. Here we analyze the meme by categorizing them as offensive or not offensive and this becomes a binary classification problem. We propose a novel offensive meme classification using the transformer-based image encoder, BiLSTM for text with mean pooling as text encoder and a Feed-Forward Network as a classification head. The SwinT + BiLSTM has performed better when compared to the ViT + BiLSTM across all the dimensions. The performance of the models has improved significantly when the contextual embeddings from DistilBert replace the custom embeddings. We have achieved the highest recall of 0.631 by combining outputs of four models using the soft voting technique

Model Architecture

Results

image

Find full discussion in the paper

Feel free to contact any authors for further discussions