/multi-modal-scale

Primary LanguagePythonMIT LicenseMIT

Hateful Meme Detection through Context-Sensitive Prompting and Fine-Grained Labeling [English | 中文]

This repository contains the codes and datasets used in our study.

Figure 1. The conceptual framework

Paper

Publication: AAAI 2025 (Student Abstract, Oral)

Authors: Rongxin Ouyang$^1$, Kokil Jaidka$^1$ $^2$, Subhayan Mukerjee$^1$ $^2$, and Guangyu Cui $^2$

$^1$ Department of Communications and New Media, National University of Singapore
$^2$ Centre for Trusted Internet & Community, National University of Singapore

Link to Paper:

  • AAAI Proceedings: [TBD] (main)
  • ArXiv: 2411.10480 (main + supplementary information)

Link to Tutorial, for social scientis who are willing to annotate images using large language models, we provided a simplified tutorial tested on Google Colab's free resources:

Dataset

Due to the size and copyright restrictions of the original dataset, please use the provided links to access the dataset for our study.

Models

We thank all contributors of the prior models used in our study:

File Structure

  • ./dataset/
    • ./dataset/raw/hateful_memes_expanded/ Meta Hateful Memes Meta Data
    • ./dataset/raw/hateful_memes_expanded/img/ Meta Hateful Memes Images
    • ...
  • ./process/
    • ./process/internvl_finetuned/ Finetuned InternVL models
    • ...
  • ./script/
    • ./script/1.finetune.distilbert.sample.ipynb Finetuning DistilBERT (unimodal)
    • ./script/2.finetune.internvl.sample.sh Finetuning Internvl 2.0 8B (multi-modal)
    • ./script/3.evaluation.batch.py Evaluations of all models
    • ...

Acknowledgment

This work was supported by the Singapore Ministry of Education AcRF TIER 3 Grant (MOE-MOET32022-0001). We gratefully acknowledge invaluable comments and discussions with Shaz Furniturewala and Jingwei Gao.

Bug Reports

  • If you encountered any questions, feel free to reach out to Rongxin (rongxin$u.nus.edu). 😄

Citation

TBD

License

MIT License