This repository contains the implementation of the MEAL model, a Multitask Efficient trAnsformer network for Laryngoscopy. The MEAL model is designed for the classification of vocal fold images and the detection of glottic landmarks and lesions. The paper will be available soon.
- Clone this repository and create a new conda environment:
git clone https://github.com/LouisDo2108/MEAL.git
conda create -n meal python=3.8.15
cd meal
conda activate meal
- Install the required packages:
pip install -r requirements.txt
-
Download the pretrained weights and data from here: google_drive
-
Unzip the downloaded files and place them under the repository folder.
-
You can start doing things like training and validating the backbone or the whole model. Please see some example scripts in run.sh
This implementation of the MEAL model draws inspiration from the YOLOv5 and SCET architectures. We would like to acknowledge the contributions of these projects.